Graph Neural Network Library for PyTorch
Project description
Documentation  Paper  Colab Notebooks and Video Tutorials  External Resources  OGB Examples
PyG (PyTorch Geometric) is a library built upon PyTorch to easily write and train Graph Neural Networks (GNNs) for a wide range of applications related to structured data.
It consists of various methods for deep learning on graphs and other irregular structures, also known as geometric deep learning, from a variety of published papers.
In addition, it consists of easytouse minibatch loaders for operating on many small and single giant graphs, multi GPUsupport, DataPipe
support, distributed graph learning via Quiver, a large number of common benchmark datasets (based on simple interfaces to create your own), the GraphGym experiment manager, and helpful transforms, both for learning on arbitrary graphs as well as on 3D meshes or point clouds.
Click here to join our Slack community!
 Library Highlights
 Quick Tour for New Users
 Architecture Overview
 Implemented GNN Models
 Installation
Library Highlights
Whether you are a machine learning researcher or firsttime user of machine learning toolkits, here are some reasons to try out PyG for machine learning on graphstructured data.
 Easytouse and unified API: All it takes is 1020 lines of code to get started with training a GNN model (see the next section for a quick tour). PyG is PyTorchontherocks: It utilizes a tensorcentric API and keeps design principles close to vanilla PyTorch. If you are already familiar with PyTorch, utilizing PyG is straightforward.
 Comprehensive and wellmaintained GNN models: Most of the stateoftheart Graph Neural Network architectures have been implemented by library developers or authors of research papers and are ready to be applied.
 Great flexibility: Existing PyG models can easily be extended for conducting your own research with GNNs. Making modifications to existing models or creating new architectures is simple, thanks to its easytouse message passing API, and a variety of operators and utility functions.
 Largescale realworld GNN models: We focus on the need of GNN applications in challenging realworld scenarios, and support learning on diverse types of graphs, including but not limited to: scalable GNNs for graphs with millions of nodes; dynamic GNNs for node predictions over time; heterogeneous GNNs with multiple node types and edge types.
 GraphGym integration: GraphGym lets users easily reproduce GNN experiments, is able to launch and analyze thousands of different GNN configurations, and is customizable by registering new modules to a GNN learning pipeline.
Quick Tour for New Users
In this quick tour, we highlight the ease of creating and training a GNN model with only a few lines of code.
Train your own GNN model
In the first glimpse of PyG, we implement the training of a GNN for classifying papers in a citation graph.
For this, we load the Cora dataset, and create a simple 2layer GCN model using the predefined GCNConv
:
import torch
from torch import Tensor
from torch_geometric.nn import GCNConv
from torch_geometric.datasets import Planetoid
dataset = Planetoid(root='.', name='Cora')
class GCN(torch.nn.Module):
def __init__(self, in_channels, hidden_channels, out_channels):
super().__init__()
self.conv1 = GCNConv(in_channels, hidden_channels)
self.conv2 = GCNConv(hidden_channels, out_channels)
def forward(self, x: Tensor, edge_index: Tensor) > Tensor:
# x: Node feature matrix of shape [num_nodes, in_channels]
# edge_index: Graph connectivity matrix of shape [2, num_edges]
x = self.conv1(x, edge_index).relu()
x = self.conv2(x, edge_index)
return x
model = GCN(dataset.num_features, 16, dataset.num_classes)
We can now optimize the model in a training loop, similar to the standard PyTorch training procedure.
import torch.nn.functional as F
data = dataset[0]
optimizer = torch.optim.Adam(model.parameters(), lr=0.01)
for epoch in range(200):
pred = model(data.x, data.edge_index)
loss = F.cross_entropy(pred[data.train_mask], data.y[data.train_mask])
# Backpropagation
optimizer.zero_grad()
loss.backward()
optimizer.step()
More information about evaluating final model performance can be found in the corresponding example.
Create your own GNN layer
In addition to the easy application of existing GNNs, PyG makes it simple to implement custom Graph Neural Networks (see here for the accompanying tutorial). For example, this is all it takes to implement the edge convolutional layer from Wang et al.:
$$x_i^{\prime} ~ = ~ \max_{j \in \mathcal{N}(i)} ~ \textrm{MLP}_{\theta} \left( [ ~ x_i, ~ x_j  x_i ~ ] \right)$$
import torch
from torch import Tensor
from torch.nn import Sequential, Linear, ReLU
from torch_geometric.nn import MessagePassing
class EdgeConv(MessagePassing):
def __init__(self, in_channels, out_channels):
super().__init__(aggr="max") # "Max" aggregation.
self.mlp = Sequential(
Linear(2 * in_channels, out_channels),
ReLU(),
Linear(out_channels, out_channels),
)
def forward(self, x: Tensor, edge_index: Tensor) > Tensor:
# x: Node feature matrix of shape [num_nodes, in_channels]
# edge_index: Graph connectivity matrix of shape [2, num_edges]
return self.propagate(edge_index, x=x) # shape [num_nodes, out_channels]
def message(self, x_j: Tensor, x_i: Tensor) > Tensor:
# x_j: Source node features of shape [num_edges, in_channels]
# x_i: Target node features of shape [num_edges, in_channels]
edge_features = torch.cat([x_i, x_j  x_i], dim=1)
return self.mlp(edge_features) # shape [num_edges, out_channels]
Manage experiments with GraphGym
GraphGym allows you to manage and launch GNN experiments, using a highly modularized pipeline (see here for the accompanying tutorial).
git clone https://github.com/pygteam/pytorch_geometric.git
cd pytorch_geometric/graphgym
bash run_single.sh # run a single GNN experiment (node/edge/graphlevel)
bash run_batch.sh # run a batch of GNN experiments, using differnt GNN designs/datasets/tasks
Users are highly encouraged to check out the documentation, which contains additional tutorials on the essential functionalities of PyG, including data handling, creation of datasets and a full list of implemented methods, transforms, and datasets.
For a quick start, check out our examples in examples/
.
Architecture Overview
PyG provides a multilayer framework that enables users to build Graph Neural Network solutions on both low and high levels. It comprises of the following components:
 The PyG engine utilizes the powerful PyTorch deep learning framework, as well as additions of efficient CUDA libraries for operating on sparse data, e.g.,
pyglib
,torchscatter
,torchsparse
andtorchcluster
.  The PyG storage handles data processing, transformation and loading pipelines. It is capable of handling and processing largescale graph datasets, and provides effective solutions for heterogeneous graphs. It further provides a variety of sampling solutions, which enable training of GNNs on largescale graphs.
 The PyG operators bundle essential functionalities for implementing Graph Neural Networks. PyG supports important GNN building blocks that can be combined and applied to various parts of a GNN model, ensuring rich flexibility of GNN design.
 Finally, PyG provides an abundant set of GNN models, and examples that showcase GNN models on standard graph benchmarks. Thanks to its flexibility, users can easily build and modify custom GNN models to fit their specific needs.
Implemented GNN Models
We list currently supported PyG models, layers and operators according to category:
GNN layers:
All Graph Neural Network layers are implemented via the nn.MessagePassing
interface.
A GNN layer specifies how to perform message passing, i.e. by designing different message, aggregation and update functions as defined here.
These GNN layers can be stacked together to create Graph Neural Network models.
 GCNConv from Kipf and Welling: SemiSupervised Classification with Graph Convolutional Networks (ICLR 2017) [Example]
 ChebConv from Defferrard et al.: Convolutional Neural Networks on Graphs with Fast Localized Spectral Filtering (NIPS 2016) [Example]
 GATConv from Veličković et al.: Graph Attention Networks (ICLR 2018) [Example]
Expand to see all implemented GNN layers...
 GCN2Conv from Chen et al.: Simple and Deep Graph Convolutional Networks (ICML 2020) [Example1, Example2]
 SplineConv from Fey et al.: SplineCNN: Fast Geometric Deep Learning with Continuous BSpline Kernels (CVPR 2018) [Example1, Example2]
 NNConv from Gilmer et al.: Neural Message Passing for Quantum Chemistry (ICML 2017) [Example1, Example2]
 CGConv from Xie and Grossman: Crystal Graph Convolutional Neural Networks for an Accurate and Interpretable Prediction of Material Properties (Physical Review Letters 120, 2018)
 ECConv from Simonovsky and Komodakis: EdgeConditioned Convolution on Graphs (CVPR 2017)
 EGConv from Tailor et al.: Adaptive Filters and Aggregator Fusion for Efficient Graph Convolutions (GNNSys 2021) [Example]
 GATv2Conv from Brody et al.: How Attentive are Graph Attention Networks? (ICLR 2022)
 TransformerConv from Shi et al.: Masked Label Prediction: Unified Message Passing Model for SemiSupervised Classification (CoRR 2020) [Example]
 SAGEConv from Hamilton et al.: Inductive Representation Learning on Large Graphs (NIPS 2017) [Example1, Example2, Example3, Example4]
 GraphConv from, e.g., Morris et al.: Weisfeiler and Leman Go Neural: Higherorder Graph Neural Networks (AAAI 2019)
 GatedGraphConv from Li et al.: Gated Graph Sequence Neural Networks (ICLR 2016)
 ResGatedGraphConv from Bresson and Laurent: Residual Gated Graph ConvNets (CoRR 2017)
 GINConv from Xu et al.: How Powerful are Graph Neural Networks? (ICLR 2019) [Example]
 GINEConv from Hu et al.: Strategies for Pretraining Graph Neural Networks (ICLR 2020)
 ARMAConv from Bianchi et al.: Graph Neural Networks with Convolutional ARMA Filters (CoRR 2019) [Example]
 SGConv from Wu et al.: Simplifying Graph Convolutional Networks (CoRR 2019) [Example]
 APPNP from Klicpera et al.: Predict then Propagate: Graph Neural Networks meet Personalized PageRank (ICLR 2019) [Example]
 MFConv from Duvenaud et al.: Convolutional Networks on Graphs for Learning Molecular Fingerprints (NIPS 2015)
 AGNNConv from Thekumparampil et al.: Attentionbased Graph Neural Network for SemiSupervised Learning (CoRR 2017) [Example]
 TAGConv from Du et al.: Topology Adaptive Graph Convolutional Networks (CoRR 2017) [Example]
 PNAConv from Corso et al.: Principal Neighbourhood Aggregation for Graph Nets (CoRR 2020) [Example]
 FAConv from Bo et al.: Beyond LowFrequency Information in Graph Convolutional Networks (AAAI 2021)
 PDNConv from Rozemberczki et al.: Pathfinder Discovery Networks for Neural Message Passing (WWW 2021)
 RGCNConv from Schlichtkrull et al.: Modeling Relational Data with Graph Convolutional Networks (ESWC 2018) [Example1, Example2]
 RGATConv from Busbridge et al.: Relational Graph Attention Networks (CoRR 2019) [Example]
 FiLMConv from Brockschmidt: GNNFiLM: Graph Neural Networks with Featurewise Linear Modulation (ICML 2020) [Example]
 SignedConv from Derr et al.: Signed Graph Convolutional Network (ICDM 2018) [Example]
 DNAConv from Fey: Just Jump: Dynamic Neighborhood Aggregation in Graph Neural Networks (ICLRW 2019) [Example]
 PANConv from Ma et al.: Path Integral Based Convolution and Pooling for Graph Neural Networks (NeurIPS 2020)
 PointConv (including Iterative Farthest Point Sampling, dynamic graph generation based on nearest neighbor or maximum distance, and kNN interpolation for upsampling) from Qi et al.: PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation (CVPR 2017) and PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space (NIPS 2017) [Example1, Example2]
 EdgeConv from Wang et al.: Dynamic Graph CNN for Learning on Point Clouds (CoRR, 2018) [Example1, Example2]
 XConv from Li et al.: PointCNN: Convolution On XTransformed Points (NeurIPS 2018) [Example]
 PPFConv from Deng et al.: PPFNet: Global Context Aware Local Features for Robust 3D Point Matching (CVPR 2018)
 GMMConv from Monti et al.: Geometric Deep Learning on Graphs and Manifolds using Mixture Model CNNs (CVPR 2017)
 FeaStConv from Verma et al.: FeaStNet: FeatureSteered Graph Convolutions for 3D Shape Analysis (CVPR 2018)
 PointTransformerConv from Zhao et al.: Point Transformer (2020)
 HypergraphConv from Bai et al.: Hypergraph Convolution and Hypergraph Attention (CoRR 2019)
 GravNetConv from Qasim et al.: Learning Representations of Irregular Particledetector Geometry with Distanceweighted Graph Networks (European Physics Journal C, 2019)
 SuperGAT from Kim and Oh: How To Find Your Friendly Neighborhood: Graph Attention Design With SelfSupervision (ICLR 2021) [Example]
 HGTConv from Hu et al.: Heterogeneous Graph Transformer (WWW 2020) [Example]
 HEATConv from Mo et al.: Heterogeneous EdgeEnhanced Graph Attention Network For MultiAgent Trajectory Prediction (CoRR 2021)
 A MetaLayer for building any kind of graph network similar to the TensorFlow Graph Nets library from Battaglia et al.: Relational Inductive Biases, Deep Learning, and Graph Networks (CoRR 2018)
 SSGConv from Zhu et al.: Simple Spectral Graph Convolution (ICLR 2021)
 FusedGATConv from Zhang et al.: Understanding GNN Computational Graph: A Coordinated Computation, IO, and Memory Perspective (MLSys 2022)
Pooling layers: Graph pooling layers combine the vectorial representations of a set of nodes in a graph (or a subgraph) into a single vector representation that summarizes its properties of nodes. It is commonly applied to graphlevel tasks, which require combining node features into a single graph representation.
 TopK Pooling from Gao and Ji: Graph UNets (ICML 2019), Cangea et al.: Towards Sparse Hierarchical Graph Classifiers (NeurIPSW 2018) and Knyazev et al.: Understanding Attention and Generalization in Graph Neural Networks (ICLRW 2019) [Example]
 DiffPool from Ying et al.: Hierarchical Graph Representation Learning with Differentiable Pooling (NeurIPS 2018) [Example]
Expand to see all implemented pooling layers...
 Attentional Aggregation from Li et al.: Graph Matching Networks for Learning the Similarity of Graph Structured Objects (ICML 2019) [Example]
 Set2Set from Vinyals et al.: Order Matters: Sequence to Sequence for Sets (ICLR 2016) [Example]
 Sort Aggregation from Zhang et al.: An EndtoEnd Deep Learning Architecture for Graph Classification (AAAI 2018) [Example]
 MinCut Pooling from Bianchi et al.: Spectral Clustering with Graph Neural Networks for Graph Pooling (ICML 2020) [Example]
 DMoN Pooling from Tsitsulin et al.: Graph Clustering with Graph Neural Networks (CoRR 2020) [Example]
 Graclus Pooling from Dhillon et al.: Weighted Graph Cuts without Eigenvectors: A Multilevel Approach (PAMI 2007) [Example]
 Voxel Grid Pooling from, e.g., Simonovsky and Komodakis: Dynamic EdgeConditioned Filters in Convolutional Neural Networks on Graphs (CVPR 2017) [Example]
 SAG Pooling from Lee et al.: SelfAttention Graph Pooling (ICML 2019) and Knyazev et al.: Understanding Attention and Generalization in Graph Neural Networks (ICLRW 2019) [Example]
 Edge Pooling from Diehl et al.: Towards Graph Pooling by Edge Contraction (ICMLW 2019) and Diehl: Edge Contraction Pooling for Graph Neural Networks (CoRR 2019) [Example]
 ASAPooling from Ranjan et al.: ASAP: Adaptive Structure Aware Pooling for Learning Hierarchical Graph Representations (AAAI 2020) [Example]
 PANPooling from Ma et al.: Path Integral Based Convolution and Pooling for Graph Neural Networks (NeurIPS 2020)
 MemPooling from Khasahmadi et al.: MemoryBased Graph Networks (ICLR 2020) [Example]
 Graph Multiset Transformer from Baek et al.: Accurate Learning of Graph Representations with Graph Multiset Pooling (ICLR 2021) [Example]
 Equilibrium Aggregation from Bartunov et al.: (UAI 2022) [Example]
GNN models: Our supported GNN models incorporate multiple message passing layers, and users can directly use these predefined models to make predictions on graphs. Unlike simple stacking of GNN layers, these models could involve preprocessing, additional learnable parameters, skip connections, graph coarsening, etc.
 SchNet from Schütt et al.: SchNet: A Continuousfilter Convolutional Neural Network for Modeling Quantum Interactions (NIPS 2017) [Example]
 DimeNet and DimeNetPlusPlus from Klicpera et al.: Directional Message Passing for Molecular Graphs (ICLR 2020) and Fast and UncertaintyAware Directional Message Passing for NonEquilibrium Molecules (NeurIPSW 2020) [Example]
 Node2Vec from Grover and Leskovec: node2vec: Scalable Feature Learning for Networks (KDD 2016) [Example]
 Deep Graph Infomax from Veličković et al.: Deep Graph Infomax (ICLR 2019) [Example1, Example2]
 Deep Multiplex Graph Infomax from Park et al.: Unsupervised Attributed Multiplex Network Embedding (AAAI 2020) [Example]
 Masked Label Prediction from Shi et al.: Masked Label Prediction: Unified Message Passing Model for SemiSupervised Classification (CoRR 2020) [Example]
Expand to see all implemented GNN models...
 Jumping Knowledge from Xu et al.: Representation Learning on Graphs with Jumping Knowledge Networks (ICML 2018) [Example]
 MetaPath2Vec from Dong et al.: metapath2vec: Scalable Representation Learning for Heterogeneous Networks (KDD 2017) [Example]
 All variants of Graph Autoencoders and Variational Autoencoders from:
 Variational Graph AutoEncoders from Kipf and Welling (NIPSW 2016) [Example]
 Adversarially Regularized Graph Autoencoder for Graph Embedding from Pan et al. (IJCAI 2018) [Example]
 Simple and Effective Graph Autoencoders with OneHop Linear Models from Salha et al. (ECML 2020) [Example]
 SEAL from Zhang and Chen: Link Prediction Based on Graph Neural Networks (NeurIPS 2018) [Example]
 RENet from Jin et al.: Recurrent Event Network for Reasoning over Temporal Knowledge Graphs (ICLRW 2019) [Example]
 GraphUNet from Gao and Ji: Graph UNets (ICML 2019) [Example]
 AttentiveFP from Xiong et al.: Pushing the Boundaries of Molecular Representation for Drug Discovery with the Graph Attention Mechanism (J. Med. Chem. 2020) [Example]
 DeepGCN and the GENConv from Li et al.: DeepGCNs: Can GCNs Go as Deep as CNNs? (ICCV 2019) and DeeperGCN: All You Need to Train Deeper GCNs (CoRR 2020) [Example]
 RECT from Wang et al.: Network Embedding with Completelyimbalanced Labels (TKDE 2020) [Example]
 GNNExplainer from Ying et al.: GNNExplainer: Generating Explanations for Graph Neural Networks (NeurIPS 2019) [Example1, Example2]
 Graphless Neural Networks from Zhang et al.: Graphless Neural Networks: Teaching Old MLPs New Tricks via Distillation (CoRR 2021) [Example]
 LINKX from Lim et al.: Large Scale Learning on NonHomophilous Graphs: New Benchmarks and Strong Simple Methods (NeurIPS 2021) [Example]
 RevGNN from Li et al.: Training Graph Neural with 1000 Layers (ICML 2021) [Example]
GNN operators and utilities: PyG comes with a rich set of neural network operators that are commonly used in many GNN models. They follow an extensible design: It is easy to apply these operators and graph utilities to existing GNN layers and models to further enhance model performance.
 DropEdge from Rong et al.: DropEdge: Towards Deep Graph Convolutional Networks on Node Classification (ICLR 2020)
 DropNode, MaskFeature and AddRandomEdge from You et al.: Graph Contrastive Learning with Augmentations (NeurIPS 2020)
 DropPath from Li et al.: MaskGAE: Masked Graph Modeling Meets Graph Autoencoders (arXiv 2022)
 ShuffleNode from Veličković et al.: Deep Graph Infomax (ICLR 2019)
 GraphNorm from Cai et al.: GraphNorm: A Principled Approach to Accelerating Graph Neural Network Training (ICML 2021)
 GDC from Klicpera et al.: Diffusion Improves Graph Learning (NeurIPS 2019) [Example]
Expand to see all implemented GNN operators and utilities...
 GraphSizeNorm from Dwivedi et al.: Benchmarking Graph Neural Networks (CoRR 2020)
 PairNorm from Zhao and Akoglu: PairNorm: Tackling Oversmoothing in GNNs (ICLR 2020)
 MeanSubtractionNorm from Yang et al.: Revisiting "Oversmoothing" in Deep GCNs (CoRR 2020)
 DiffGroupNorm from Zhou et al.: Towards Deeper Graph Neural Networks with Differentiable Group Normalization (NeurIPS 2020)
 Tree Decomposition from Jin et al.: Junction Tree Variational Autoencoder for Molecular Graph Generation (ICML 2018)
 TGN from Rossi et al.: Temporal Graph Networks for Deep Learning on Dynamic Graphs (GRL+ 2020) [Example]
 Weisfeiler Lehman Operator from Weisfeiler and Lehman: A Reduction of a Graph to a Canonical Form and an Algebra Arising During this Reduction (NauchnoTechnicheskaya Informatsia 1968) [Example]
 Continuous Weisfeiler Lehman Operator from Togninalli et al.: Wasserstein WeisfeilerLehman Graph Kernels (NeurIPS 2019)
 Label Propagation from Zhu and Ghahramani: Learning from Labeled and Unlabeled Data with Label Propagation (CMUCALD 2002) [Example]
 Local Degree Profile from Cai and Wang: A Simple yet Effective Baseline for Nonattribute Graph Classification (CoRR 2018)
 CorrectAndSmooth from Huang et al.: Combining Label Propagation And Simple Models Outperforms Graph Neural Networks (CoRR 2020) [Example]
 Gini and BRO regularization from Henderson et al.: Improving Molecular Graph Neural Network Explainability with Orthonormalization and Induced Sparsity (ICML 2021)
 RootedEgoNets and RootedRWSubgraph from Zhao et al.: From Stars to Subgraphs: Uplifting Any GNN with Local Structure Awareness (ICLR 2022)
 FeaturePropagation from Rossi et al.: On the Unreasonable Effectiveness of Feature Propagation in Learning on Graphs with Missing Node Features (CoRR 2021)
Scalable GNNs: PyG supports the implementation of Graph Neural Networks that can scale to largescale graphs. Such application is challenging since the entire graph, its associated features and the GNN parameters cannot fit into GPU memory. Many stateoftheart scalability approaches tackle this challenge by sampling neighborhoods for minibatch training, graph clustering and partitioning, or by using simplified GNN models. These approaches have been implemented in PyG, and can benefit from the above GNN layers, operators and models.
 NeighborSampler and NeighborLoader from Hamilton et al.: Inductive Representation Learning on Large Graphs (NIPS 2017) [Example1, Example2, Example3, Example4]
 ClusterGCN from Chiang et al.: ClusterGCN: An Efficient Algorithm for Training Deep and Large Graph Convolutional Networks (KDD 2019) [Example1, Example2]
 GraphSAINT from Zeng et al.: GraphSAINT: Graph Sampling Based Inductive Learning Method (ICLR 2020) [Example]
Expand to see all implemented scalable GNNs...
 ShaDow from Zeng et al.: Decoupling the Depth and Scope of Graph Neural Networks (NeurIPS 2021) [Example]
 SIGN from Rossi et al.: SIGN: Scalable Inception Graph Neural Networks (CoRR 2020) [Example]
 HGTLoader from Hu et al.: Heterogeneous Graph Transformer (WWW 2020) [Example]
Installation
PyG is available for Python 3.7 to Python 3.10.
Anaconda
Update: You can now install PyG via Anaconda for all major OS/PyTorch/CUDA combinations 🤗 Given that you have PyTorch >= 1.8.0 installed, simply run
conda install pyg c pyg
Pip Wheels
We alternatively provide pip wheels for all major OS/PyTorch/CUDA combinations, see here.
PyTorch 1.13
To install the binaries for PyTorch 1.13.0, simply run
pip install pyglib torchscatter torchsparse f https://data.pyg.org/whl/torch1.13.0+${CUDA}.html
pip install torchgeometric
where ${CUDA}
should be replaced by either cpu
, cu116
, or cu117
depending on your PyTorch installation.
cpu 
cu116 
cu117 


Linux  ✅  ✅  ✅ 
Windows  ✅  ✅  ✅ 
macOS  ✅ 
For additional but optional functionality, run
pip install torchcluster torchsplineconv f https://data.pyg.org/whl/torch1.13.0+${CUDA}.html
PyTorch 1.12
To install the binaries for PyTorch 1.12.0, simply run
pip install pyglib torchscatter torchsparse f https://data.pyg.org/whl/torch1.12.0+${CUDA}.html
pip install torchgeometric
where ${CUDA}
should be replaced by either cpu
, cu102
, cu113
, or cu116
depending on your PyTorch installation.
cpu 
cu102 
cu113 
cu116 


Linux  ✅  ✅  ✅  ✅ 
Windows  ✅  ✅  ✅  
macOS  ✅ 
For additional but optional functionality, run
pip install torchcluster torchsplineconv f https://data.pyg.org/whl/torch1.12.0+${CUDA}.html
Note: Binaries of older versions are also provided for PyTorch 1.4.0, PyTorch 1.5.0, PyTorch 1.6.0, PyTorch 1.7.0/1.7.1, PyTorch 1.8.0/1.8.1, PyTorch 1.9.0, PyTorch 1.10.0/1.10.1/1.10.2, and PyTorch 1.11.0 (following the same procedure).
For older versions, you might need to explicitly specify the latest supported version number or install via pip install noindex
in order to prevent a manual installation from source.
You can look up the latest supported version number here.
Nightly and Master
In case you want to experiment with the latest PyG features which are not fully released yet, ensure that pyglib
, torchscatter
and torchsparse
are installed by following the steps mentioned above, and install either the nightly version of PyG via
pip install pygnightly
or install PyG from master via
pip install git+https://github.com/pygteam/pytorch_geometric.git
Cite
Please cite our paper (and the respective papers of the methods used) if you use this code in your own work:
@inproceedings{Fey/Lenssen/2019,
title={Fast Graph Representation Learning with {PyTorch Geometric}},
author={Fey, Matthias and Lenssen, Jan E.},
booktitle={ICLR Workshop on Representation Learning on Graphs and Manifolds},
year={2019},
}
Feel free to email us if you wish your work to be listed in the external resources. If you notice anything unexpected, please open an issue and let us know. If you have any questions or are missing a specific feature, feel free to discuss them with us. We are motivated to constantly make PyG even better.
Project details
Release history Release notifications  RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.