📘Documentation | 🛠️Installation | 🚀Model Zoo | 🆕News
This repository is an open-source project for video prediction benchmarks, which contains the implementation code for paper:
SimVP: Towards Simple yet Powerful Spatiotemporal Predictive learning
Cheng Tan, Zhangyang Gao, Siyuan Li, Stan Z. Li.
This is the journal version of our previous conference work (SimVP: Simpler yet Better Video Prediction, In CVPR 2022). It is worth noticing that the hidden Translator token mixing
and channel mixing
).
openstl/api
contains an experiment runner.openstl/core
contains core training plugins and metrics.openstl/datasets
contains datasets and dataloaders.openstl/methods/
contains training methods for various video prediction methods.openstl/models/
contains the main network architectures of various video prediction methods.openstl/modules/
contains network modules and layers.tools/train.py
andtools/train.py
are the executable python file with possible arguments for training, validating, and testing pipelines.
[2023-04-19] OpenSTL
v0.2.0 is released. The training loop and dataloaders are fixed.
This project has provided an environment setting file of conda, users can easily reproduce the environment by the following commands:
git clone https://github.com/chengtan9907/OpenSTL
cd OpenSTL
conda env create -f environment.yml
conda activate OpenSTL
python setup.py develop
Dependencies
- argparse
- fvcore
- numpy
- hickle
- scikit-image
- scikit-learn
- torch
- timm
- tqdm
- xarray==0.19.0
Please refer to install.md for more detailed instructions.
Please see get_started.md for the basic usage. Here is an example of single GPU non-distributed training SimVP+gSTA on Moving MNIST dataset.
bash tools/prepare_data/download_mmnist.sh
python tools/train.py -d mmnist --lr 1e-3 -c configs/mmnist/simvp/SimVP_gSTA.py --ex_name mmnist_simvp_gsta
We support various video prediction methods and will provide benchmarks on various video prediction datasets. We are working on add new methods and collecting experiment results.
-
Video Prediction Methods.
Currently supported methods
Currently supported MetaFormer models for SimVP
- ViT (ICLR'2021)
- Swin-Transformer (ICCV'2021)
- MLP-Mixer (NIPS'2021)
- ConvMixer (Openreview'2021)
- UniFormer (ICLR'2022)
- PoolFormer (CVPR'2022)
- ConvNeXt (CVPR'2022)
- VAN (ArXiv'2022)
- IncepU (SimVP.V1) (CVPR'2022)
- gSTA (SimVP.V2) (ArXiv'2022)
- HorNet (NIPS'2022)
- MogaNet (ArXiv'2022)
-
Video Prediction Benchmarks.
Currently supported datasets
- KTH Action (ICPR'2004) [download] [config]
- KittiCaltech Pedestrian (IJRR'2013) [download] [config]
- Moving MNIST (ICML'2015) [download] [config]
- TaxiBJ (AAAI'2017) [download] [config]
- WeatherBench (ArXiv'2020) [download] [config]
This project is released under the Apache 2.0 license. See LICENSE
for more information.
SimVPv2 is an open-source project for video prediction methods created by researchers in CAIRI AI Lab. We encourage researchers interested in video and weather prediction to contribute to SimVPv2! We borrow the official implementations of ConvLSTM, PredRNN variants, E3D-LSTM, MAU, CrevNet, and PhyDNet.
If you are interested in our repository or our paper, please cite the following paper:
@article{tan2022simvp,
title={SimVP: Towards Simple yet Powerful Spatiotemporal Predictive Learning},
author={Tan, Cheng and Gao, Zhangyang and Li, Siyuan and Li, Stan Z},
journal={arXiv preprint arXiv:2211.12509},
year={2022}
}
@misc{tan2023openstl,
title={OpenSTL: Open-source Toolbox for SpatioTemporal Predictive Learning},
author={Tan, Cheng and Li, Siyuan and Gao, Zhangyang and Li, Stan Z},
howpublished = {\url{https://github.com/chengtan9907/OpenSTL}},
year={2023}
}
For adding new features, looking for helps, or reporting bugs associated with OpenSTL
, please open a GitHub issue and pull request with the tag "help wanted" or "enhancement". Feel free to contact us through email if you have any questions. Enjoy!
- Siyuan Li (lisiyuan@westlake.edu.cn), Westlake University & Zhejiang University
- Cheng Tan (tancheng@westlake.edu.cn), Westlake University & Zhejiang University
- Zhangyang Gao (gaozhangyang@westlake.edu.cn), Westlake University & Zhejiang University