Skip to content

Implementation of Time-aware Random Walk Diffusion to Improve Dynamic Graph Learning (AAAI-23)

Notifications You must be signed in to change notification settings

Mitchell-xiyunfeng/TiaRa

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

9 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

TiaRa

This is the official implementation of TiaRa (Time-aware Random Walk Diffusion), which is described in the following paper:

  • Time-aware Random Walk Diffusion to Improve Dynamic Graph Learning
    Jong-whi Lee and Jinhong Jung
    37-th AAAI Conference on Artificial Intelligence (AAAI) 2023, Washington, DC, USA

The paper is publicly accessible in the following link:

Overview

How can we augment a dynamic graph for improving the performance of dynamic graph neural networks?

Graph augmentation has been widely utilized to boost the learning performance of GNN-based models. However, most existing approaches only enhance spatial structure within an input static graph by transforming the graph, and do not consider dynamics caused by time such as temporal locality, i.e., recent edges are more influential than earlier ones, which remains challenging for dynamic graph augmentation.

In this work, we propose TiaRa (Time-aware Random Walk Diffusion), a novel diffusion-based method for augmenting a dynamic graph represented as a discrete-time sequence of graph snapshots. For this purpose, we first design a time-aware random walk proximity so that a surfer can walk along the time dimension as well as edges, resulting in spatially and temporally localized scores. We then derive our diffusion matrices based on the time-aware random walk, and show they become enhanced adjacency matrices that both spatial and temporal localities are augmented.

Prerequisites

This implementation has been tested in conda virtual environment. Please run conda env create -f environment.yml [-n ENVNAME] to create it. Note that the default name of the environment is tiara. Please see the full list of packages to be insalled in environment.yml where you can change their versions if necessary. The representative packages that we use are as follows:

  • pytorch
  • torchmetrics
  • pytorch-sparse
  • dgl
  • pytorch-geometric
  • pytorch-geometric-temporal

Datasets and Settings

We provide datasets and preprocessing code used in the paper. The current ./data contains datasets only for node classification, but datasets for temporal link prediction will be automatically downloaded at runtime.

Dataset Nodes Edges Time Step Features Labels Task
BitcoinAlpha 3,783 31,748 138 32 2 Link
WikiElec 7,125 212,854 100 32 2 Link
RedditBody 35,776 484,460 88 32 2 Link
Brain 5,000 1,955,488 12 20 10 Node
DBLP3 4,257 23,540 10 100 3 Node
DBLP5 6,606 42,815 10 100 5 Node
Reddit 8,291 264,050 10 20 4 Node

We perform experiments with a list of random seeds, {117, 3690, 2534, 1576, 1781}, and the searched hyperparameters of all models with TiaRa and datasets are at settings. We conducted our experiments on RTX 3090 (24GB VRAM) with CUDA 11.3.

Demo

We included a demo script link_demo.sh which reproduces our experiments for the temporal link prediction task.

bash link_demo.sh

The above script runs an experiment of GCRN+TiaRa on the BitcoinAlpha dataset with a random seed 117 where the searched hyperparameters are found at settings/GCRN-BitcoinAlpha-tiara.json.

The follofing script reproduces our experiments for the node classification task.

bash node_demo.sh

The script runs an experiment of GCRN+TiaRa on the Brain dataset with a random seed 117 where the searched hyperparameters are found at settings/GCRN-Brain-tiara.json.

If you want to perform an experiment on another dataset, use run.sh:

bash run.sh ${DATASET} ${MODEL} ${SEED} [arguments]

where ${DATASET} is in the dataset table and ${MODEL} is GCRN, EvolveGCN, and GCN. You can change [arguments] which are options described below if necessary.

Usage and Options

The run.sh uses src/main.py to conduct experiments with TiaRa where the implmentation of TiaRa is found at src/augmenter.py.

python src/main.py [--<argument name> <argument value>] [...]

We describe the detailed options of src/main.py. The following table summarizes options related to device, seed, and data.

Option Description Default
device device name cuda
seed random seed for reproduction None
dataset dataset name BitcoinAlpha
time-aggregation length of time range in a single time step 1200000
train-ratio ratio for train split 0.7
val-ratio ratio for validation split 0.1
data-dir dataset path data
verbose print additional informations False
  • If your GPU memory is small, then you might encounter CUDA out of memory error. In this case, consider --device cpu on a workstation with enough memory space.

The following table summarizes options related to augmentation methods.

Option Description Default
augment-method graph augmentation method tiara
alpha restart probability for tiara 0.2
beta time travel probability for tiara 0.3
eps filtering threshold for tiara 1e-3
K number of power iteration for tiara 100
symmetric-trick symmetric trick strategy for tiara True
dropedge dropedge ratio 0

The following table summarizes options related to optimizers where we use Adam optimizer in this work.

Option Description Default
lr learning rate 0.05
weight-decay weight decay value 0.0001
lr-decay learning rate decay value 0.999

The following table summarizes options related to GNN models.

Option Description Default
model GNN model name GCRN
input-dim input dimmension size for GNN model 32
hidden-dim hidden dimmension size for GNN model 32
output-dim outoput dimmension size for GNN model 32
decoder-dim hidden dimmension size for decoder model 32
num-blocks number of layers for GNN model 3
rnn RNN model name for GNN model except GCN LSTM
dropout dropout ratio 0
  • model: {GCRN, EvolveGCN, GCN}
  • rnn: {LSTM, GRU}

How to use TiaRa in My Code

TiaRa class in src/augmenter.py contains the implementation of our proposed method. You may import or copy the TiaRa class in your work. Each augmenter including TiaRa takes its parameters at initialization. The dynamic graph (or a list of graph snapshot) is augmented by calling the augmenter object. Each graph snapshot in the dynamic graph should be a dgl object. You may refer to our implementation in src/dataset_loader in order to check how it preprocesses raw datasets.

We also provide the usage of TiaRa in the following example:

# `original_dynamic_graph` is a list of graph snapshots where each snapshot is a dgl object

tiara = TiaRa(alpha=0.2, beta=0.3, eps=1e-3, K=100, device='cuda')
augmented_dynamic_graph = tiara(original_dynamic_graph)

# `augmented_dynamic_graph` is also a list of (augmented) graph snapshots, and fed to a dynamic GNN model

Information of other implementations

We refer to open-source implementation of GNN models and augmentation methods at the following links:

Citation

Please cite the paper if you use this code in your own work:

@inproceedings{LeeJ2023tiara,
  title={Time-aware Random Walk Diffusion to Improve Dynamic Graph Learning},
  author={Jong{-}whi Lee and Jinhong Jung},
  booktitle={Proceedings of the AAAI Conference on Artificial Intelligence},
  year={2023}
}

About

Implementation of Time-aware Random Walk Diffusion to Improve Dynamic Graph Learning (AAAI-23)

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 99.6%
  • Shell 0.4%