This repo is the open-sourced code for our work TGL: A General Framework for Temporal Graph Training on Billion-Scale Graphs.
- python >= 3.6.13
- pytorch >= 1.8.1
- pandas >= 1.1.5
- numpy >= 1.19.5
- dgl >= 0.6.1
- pyyaml >= 5.4.1
- tqdm >= 4.61.0
- pybind11 >= 2.6.2
- g++ >= 7.5.0
- openmp >= 201511
Our temporal sampler is implemented using C++, please compile the sampler first with the following command
python setup.py build_ext --inplace
The four datasets used in our paper are available to download from AWS S3 bucket using the down.sh
script. The total download size is around 350GB.
To use your own dataset, you need to put the following files in the folder \DATA\\<NameOfYourDataset>\
edges.csv
: The file that stores temporal edge informations. The csv should have the following columns with the header as,src,dst,time,ext_roll
where each of the column refers to edge index (start with zero), source node index (start with zero), destination node index, time stamp, extrapolation roll (0 for training edges, 1 for validation edges, 2 for test edges). The CSV should be sorted by time ascendingly.ext_full.npz
: The T-CSR representation of the temporal graph. We provide a script to generate this file fromedges.csv
. You can use the following command to use the scriptpython gen_graph.py --data <NameOfYourDataset>
edge_features.pt
(optional): The torch tensor that stores the edge featrues row-wise with shape (num edges, dim edge features). Note: at least one ofedge_features.pt
ornode_features.pt
should present.node_features.pt
(optional): The torch tensor that stores the node featrues row-wise with shape (num nodes, dim node features). Note: at least one ofedge_features.pt
ornode_features.pt
should present.labels.csv
(optional): The file contains node labels for dynamic node classification task. The csv should have the following columns with the header as,node,time,label,ext_roll
where each of the column refers to node label index (start with zero), node index (start with zero), time stamp, node label, extrapolation roll (0 for training node labels, 1 for validation node labels, 2 for test node labels). The CSV should be sorted by time ascendingly.
We provide example configuration files for five temporal GNN methods: JODIE, DySAT, TGAT, TGN and TGAT. The configuration files for single GPU training are located at /config/
while the multiple GPUs training configuration files are located at /config/dist/
.
The provided configuration files are all tested to be working. If you want to use your own network architecture, please refer to /config/readme.yml
for the meaining of each entry in the yaml configuration file. As our framework is still under development, it possible that some combination of the confiruations will lead to bug.
Currently, our framework only supports extrapolation setting (inference for the future).
python train.py --data <NameOfYourDataset> --config <PathToConfigFile>
python -m torch.distributed.launch --nproc_per_node=<NumberOfGPUs+1> train_dist.py --data <NameOfYourDataset> --config <PathToConfigFile> --num_gpus <NumberOfGPUs>
Currenlty, TGL only supports performing dynamic node classification using the dynamic node embedding generated in link prediction.
For Single GPU models, directly run
python train_node.py --data <NameOfYourDATA> --config <PathToConfigFile> --model <PathToSavedModel>
For multi-GPU models, you need to first generate the dynamic node embedding
python -m torch.distributed.launch --nproc_per_node=<NumberOfGPUs+1> extract_node_dist.py --data <NameOfYourDataset> --config <PathToConfigFile> --num_gpus <NumberOfGPUs> --model <PathToSavedModel>
After generating the node embeding for multi-GPU models, run
python train_node.py --data <NameOfYourDATA> --model <PathToSavedModel>
See CONTRIBUTING for more information.
This project is licensed under the Apache-2.0 License.