This is a collection of state-of-the-art deep model for estimating optical flow. The main goal is to provide a unified framework where multiple models can be trained and tested more easily.
The work and code from many others are present here. I tried to make sure everything is properly referenced, but please let me know if I missed something.
This is still under development, so some things may not work as intended. I plan to add more models in the future, as well keep improving the platform.
- Added new models:
- MemFlow https://arxiv.org/abs/2404.04808
- NeuFlow https://arxiv.org/abs/2403.10425
- SEA-RAFT https://arxiv.org/abs/2405.14793
- SplatFlow https://arxiv.org/abs/2306.08887
- Add support for more datasets:
- Add ONNX and TensorRT conversion to RAPIDFlow
- Fix LR scheduler when accumulating gradients
- Added new models:
- Enable FP16 in most models.
- Except the following models, since they have operations that cannot run in FP16: lcv_raft, matchflow, and separableflow
- Add FP16 mode in infer, model_benchmark, and validate scripts
- Create plot_results.py script
- Move resize operations to CUDA (thanks to coca-huang)
- Added new models:
- DIP https://arxiv.org/abs/2204.00330
- Flow1D https://arxiv.org/abs/2103.04524
- FlowFormer++ https://arxiv.org/abs/2303.01237
- GMFlow+, UniMatch https://arxiv.org/abs/2211.05783
- MatchFlow https://arxiv.org/abs/2303.08384
- MS-RAFT+ https://arxiv.org/abs/2210.16900
- RPKNet https://hmorimitsu.com/publication/2024-aaai-rpknet
- SeparableFlow https://openaccess.thecvf.com/content/ICCV2021/papers/Zhang_Separable_Flow_Learning_Motion_Cost_Volumes_for_Optical_Flow_Estimation_ICCV_2021_paper.pdf
- SKFlow https://arxiv.org/abs/2205.14623
- VideoFlow https://arxiv.org/abs/2303.08340
speed_benchmark.py
becomesmodel_benchmark.py
and records more metrics- Fix compatibility with PyTorch 2.0
- Fix compatibility with PyTorch Lightning 1.9
- Fix resizing augmentation when the valid mask is sparse
- Add support for more datasets:
- CCMR https://arxiv.org/abs/2311.02661
- CRAFT https://arxiv.org/abs/2203.16896
- CSFlow https://arxiv.org/abs/2202.00909
- DICL-Flow https://arxiv.org/abs/2010.14851
- DIP https://arxiv.org/abs/2204.00330
- FastFlowNet https://arxiv.org/abs/2103.04524
- Flow1D https://arxiv.org/abs/2103.04524
- FlowFormer https://arxiv.org/abs/2203.16194
- FlowFormer++ https://arxiv.org/abs/2303.01237
- FlowNet https://arxiv.org/abs/1504.06852
- FlowNet2 https://arxiv.org/abs/1612.01925
- GMA https://arxiv.org/abs/2104.02409
- GMFlow https://arxiv.org/abs/2111.13680
- GMFlow+, UniMatch https://arxiv.org/abs/2211.05783
- GMFlowNet https://arxiv.org/abs/2203.11335
- HD3 https://arxiv.org/abs/1812.06264
- IRR https://arxiv.org/abs/1904.05290
- LCV https://arxiv.org/abs/2007.11431
- LiteFlowNet https://arxiv.org/abs/1805.07036
- LiteFlowNet2 https://arxiv.org/abs/1903.07414
- LiteFlowNet3 https://arxiv.org/abs/2007.09319
- LLA-Flow https://arxiv.org/abs/2304.08101
- MaskFlownet https://arxiv.org/abs/2003.10955
- MatchFlow https://arxiv.org/abs/2303.08384
- MemFlow https://arxiv.org/abs/2404.04808
- MS-RAFT+ https://arxiv.org/abs/2210.16900
- NeuFlow v1 https://arxiv.org/abs/2403.10425
- NeuFlow v2 https://arxiv.org/abs/2408.10161
- PWCNet https://arxiv.org/abs/1709.02371
- RAFT https://arxiv.org/abs/2003.12039
- RAPIDFlow https://hmorimitsu.com/publication/2024-icra-rapidflow/
- RPKNet https://hmorimitsu.com/publication/2024-aaai-rpknet
- ScopeFlow https://arxiv.org/abs/2002.10770
- SCV https://arxiv.org/abs/2104.02166
- SEA-RAFT https://arxiv.org/abs/2405.14793
- SeparableFlow https://openaccess.thecvf.com/content/ICCV2021/papers/Zhang_Separable_Flow_Learning_Motion_Cost_Volumes_for_Optical_Flow_Estimation_ICCV_2021_paper.pdf
- SKFlow https://arxiv.org/abs/2205.14623
- SplatFlow https://arxiv.org/abs/2306.08887
- STaRFlow https://arxiv.org/abs/2007.05481
- VCN https://papers.nips.cc/paper/2019/file/bbf94b34eb32268ada57a3be5062fe7d-Paper.pdf
- VideoFlow https://arxiv.org/abs/2303.08340
Read more details about the models on https://ptlflow.readthedocs.io/en/latest/models/models_list.html.
You can see a table with main evaluation results of the available models here. More results are also available in the folder docs/source/results.
Disclaimer: These results are the ones obtained by evaluating the available models in this framework in my machine. Your results may be different due to differences in hardware and software. I also do not guarantee that the results of each model will be similar to the ones presented in the respective papers or other original sources. If you need to replicate the original results from a paper, you should use the original implementations.
Please take a look at the documentation to learn how to install and use PTLFlow.
You can also check the notebooks below running on Google Colab for some practical examples:
The original code of this repository is licensed under the Apache 2.0 license.
Each model may be subjected to different licenses. The license of each model is included in their respective folders. It is your responsibility to make sure that your project is in compliance with all the licenses and conditions involved.
The external pretrained weights all have different licenses, which are listed in their respective folders.
The pretrained weights that were trained within this project are available under the CC BY-NC-SA 4.0 license, which I believe that covers the licenses of the datasets used in the training. That being said, I am not a legal expert so if you plan to use them to any purpose other than research, you should check all the involved licenses by yourself. Additionally, the datasets used for the training usually require the user to cite the original papers, so be sure to include their respective references in your work.
Contribution are welcome! Please check CONTRIBUTING.md to see how to contribute.
@misc{morimitsu2021ptlflow,
author = {Henrique Morimitsu},
title = {PyTorch Lightning Optical Flow},
year = {2021},
publisher = {GitHub},
journal = {GitHub repository},
howpublished = {\url{https://github.com/hmorimitsu/ptlflow}}
}
- This README file is heavily inspired by the one from the timm repository.
- Some parts of the code were inspired by or taken from FlowNetPytorch.
- flownet2-pytorch was also another important source.
- The current main training routine is based on RAFT.