EasyTPP
is an easy-to-use development and application toolkit for Temporal Point Process (TPP), with key features in configurability, compatibility and reproducibility. We hope this project could benefit both researchers and practitioners with the goal of easily customized development and open benchmarking in TPP.
| Features | Model List | Dataset | Quick Start | Benchmark |Documentation |Todo List | Citation |Acknowledgement | Star History |
- [09-02-2023] We published two non-anthropogenic datasets earthquake and volcano eruption! See Dataset for details.
- [06-22-2023] Our paper Language Model Can Improve Event Prediction by Few-Shot Abductive Reasoning was accepted by the Knowledge and Logical Reasoning Workshop, ICML'2023!
- [05-29-2023] We released
EasyTPP
v0.0.1! - [12-27-2022] Our paper Bellman Meets Hawkes: Model-Based Reinforcement Learning via Temporal Point Processes was accepted by AAAI'2023!
- [10-01-2022] Our paper HYPRO: A Hybridly Normalized Probabilistic Model for Long-Horizon Prediction of Event Sequences was accepted by NeurIPS'2022!
- [05-01-2022] We started to develop
EasyTPP
.
Features [Back to Top]
- Configurable and customizable: models are modularized and configurable,with abstract classes to support developing customized TPP models.
- Compatible with both Tensorflow and PyTorch framework:
EasyTPP
implements two equivalent sets of models, which can be run under Tensorflow (both Tensorflow 1.13.1 and Tensorflow 2.0) and PyTorch 1.7.0+ respectively. While the PyTorch models are more popular among researchers, the compatibility with Tensorflow is important for industrial practitioners. - Reproducible: all the benchmarks can be easily reproduced.
- Hyper-parameter optimization: a pipeline of optuna-based HPO is provided.
Model List [Back to Top]
We provide reference implementations of various state-of-the-art TPP papers:
No | Publication | Model | Paper | Implementation |
---|---|---|---|---|
1 | KDD'16 | RMTPP | Recurrent Marked Temporal Point Processes: Embedding Event History to Vector | Tensorflow Torch |
2 | NeurIPS'17 | NHP | The Neural Hawkes Process: A Neurally Self-Modulating Multivariate Point Process | Tensorflow Torch |
3 | NeurIPS'19 | FullyNN | Fully Neural Network based Model for General Temporal Point Processes | Tensorflow Torch |
4 | ICML'20 | SAHP | Self-Attentive Hawkes process | Tensorflow Torch |
5 | ICML'20 | THP | Transformer Hawkes process | Tensorflow Torch |
6 | ICLR'20 | IntensityFree | Intensity-Free Learning of Temporal Point Processes | Tensorflow Torch |
7 | ICLR'21 | ODETPP | Neural Spatio-Temporal Point Processes (simplified) | Tensorflow Torch |
8 | ICLR'22 | AttNHP | Transformer Embeddings of Irregularly Spaced Events and Their Participants | Tensorflow Torch |
Dataset [Back to Top]
We preprocessed one synthetic and five real world datasets from widely-cited works that contain diverse characteristics in terms of their application domains and temporal statistics:
- Synthetic: a univariate Hawkes process simulated by Tick library.
- Retweet (Zhou, 2013): timestamped user retweet events.
- Taxi (Whong, 2014): timestamped taxi pick-up events.
- StackOverflow (Leskovec, 2014): timestamped user badge reward events in StackOverflow.
- Taobao (Xue et al, 2022): timestamped user online shopping behavior events in Taobao platform.
- Amazon (Xue et al, 2022): timestamped user online shopping behavior events in Amazon platform.
Per users' request, we processed two non-anthropogenic datasets
-
Earthquake: timestamped earthquake events over the Conterminous U.S from 1996 to 2023, processed from USGS.
-
Volcano eruption: timestamped volcano eruption events over the world in recent hundreds of years, processed from The Smithsonian Institution.
All datasets are preprocess to the
Gatech
format dataset widely used for TPP researchers, and saved at Google Drive with a public access.
Quick Start [Back to Top]
We provide an end-to-end example for users to run a standard TPP model with EasyTPP
.
First of all, we can install the package either by using pip or from the source code on Github.
To install the latest stable version:
pip install easy-tpp
To install the latest on GitHub:
git clone https://github.com/ant-research/EasyTemporalPointProcess.git
cd EasyTemporalPointProcess
python setup.py install
We need to put the datasets in a local directory before running a model and the datasets should follow a certain format. See OnlineDoc - Datasets for more details.
Suppose we use the taxi dataset in the example.
Before start training, we need to set up the config file for the pipeline. We provide a preset config file in Example Config. The details of the configuration can be found in OnlineDoc - Training Pipeline.
After the setup of data and config, the directory structure is as follows:
data
|______taxi
|____ train.pkl
|____ dev.pkl
|____ test.pkl
configs
|______experiment_config.yaml
Then we start the training by simply running the script
import argparse
from easy_tpp.config_factory import Config
from easy_tpp.runner import Runner
def main():
parser = argparse.ArgumentParser()
parser.add_argument('--config_dir', type=str, required=False, default='configs/experiment_config.yaml',
help='Dir of configuration yaml to train and evaluate the model.')
parser.add_argument('--experiment_id', type=str, required=False, default='NHP_train',
help='Experiment id in the config file.')
args = parser.parse_args()
config = Config.build_from_yaml_file(args.config_dir, experiment_id=args.experiment_id)
model_runner = Runner.build_from_config(config)
model_runner.run()
if __name__ == '__main__':
main()
A more detailed example can be found at OnlineDoc - QuickStart.
Documentation [Back to Top]
The classes and methods of EasyTPP
have been well documented so that users can generate the documentation by:
cd doc
pip install -r requirements.txt
make html
NOTE:
- The
doc/requirements.txt
is only for documentation by Sphinx, which can be automatically generated by Github actions.github/workflows/docs.yml
. (Trigger by pull request.)
The full documentation is available on the website.
Benchmark [Back to Top]
In the examples folder, we provide a script to benchmark the TPPs, with Taxi dataset as the input.
To run the script, one should download the Taxi data following the above instructions. The config file is readily setup up. Then run
cd examples
python benchmark_script.py
License [Back to Top]
This project is licensed under the Apache License (Version 2.0). This toolkit also contains some code modified from other repos under other open-source licenses. See the NOTICE file for more information.
Todo List [Back to Top]
- New dataset:
- New model:
- Meta Temporal Point Process, ICLR 2023.
- Model-based RL via TPP, AAAI 2022.
Citation [Back to Top]
If you find EasyTPP
useful for your research or development, please cite the following paper:
@article{xue2023easytpp,
title={EasyTPP: Towards Open Benchmarking the Temporal Point Processes},
author={Siqiao Xue and Xiaoming Shi and Zhixuan Chu and Yan Wang and Fan Zhou and Hongyan Hao and Caigao Jiang and Chen Pan and Yi Xu and James Y. Zhang and Qingsong Wen and Jun Zhou and Hongyuan Mei},
year={2023},
eprint={2307.08097},
archivePrefix={arXiv},
primaryClass={cs.LG}
}
Acknowledgment [Back to Top]
The project is jointly initiated by Machine Intelligence Group, Alipay and DAMO Academy, Alibaba.
The following repositories are used in EasyTPP
, either in close to original form or as an inspiration:
- EasyRec
- EasyNLP
- FuxiCTR
- Neural Hawkes Process
- Neural Hawkes Particle Smoothing
- Attentive Neural Hawkes Process
- Huggingface - transformers