# PyTorch Lightning Optical Flow
![GitHub CI python status](https://github.com/hmorimitsu/ptlflow/actions/workflows/python.yml/badge.svg)
![GitHub CI pytorch status](https://github.com/hmorimitsu/ptlflow/actions/workflows/pytorch.yml/badge.svg)
![GitHub CI lightning status](https://github.com/hmorimitsu/ptlflow/actions/workflows/lightning.yml/badge.svg)
![GitHub CI build status](https://github.com/hmorimitsu/ptlflow/actions/workflows/build.yml/badge.svg)
## Introduction
This is a collection of state-of-the-art deep model for estimating optical flow. The main goal is to provide a unified framework where multiple models can be trained and tested more easily.
The work and code from many others are present here. I tried to make sure everything is properly referenced, but please let me know if I missed something.
This is still under development, so some things may not work as intended. I plan to add more models in the future, as well keep improving the platform.
- [What's new](#whats-new)
- [Available models](#available-models)
- [Results](#results)
- [Getting started](#getting-started)
- [Licenses](#licenses)
- [Contributing](#contributing)
- [Citing](#citing)
- [Acknowledgements](#acknowledgements)
## What's new
### - v0.4.0
Major update to support Lightning 2 (finally!). However, it also introduces breaking changes from the previous v0.3 code. See the details below.
- Transitioning from v0.3 to v0.4: check the [v0.4 upgrade guide](https://ptlflow.readthedocs.io/en/latest/starting/v04_upgrade_guide.html)
- Added features:
- Support for YAML config files. See the [config file documentation](https://ptlflow.readthedocs.io/en/latest/starting/config_files.html)
- Table [comparing PTLFlow results with the original papers](https://ptlflow.readthedocs.io/en/latest/results/paper_ptlflow.html) to check the stability of the included models.
- Added new models:
- NeuFlow v2 [https://arxiv.org/abs/2408.10161](https://arxiv.org/abs/2408.10161)
- Add support for more datasets:
- Middlebury-ST [https://vision.middlebury.edu/stereo/data/scenes2014/]{https://vision.middlebury.edu/stereo/data/scenes2014/}
- VIPER [https://playing-for-benchmarks.org/](https://playing-for-benchmarks.org/)
### - v0.3.2
- Added new models:
- MemFlow [https://arxiv.org/abs/2404.04808](https://arxiv.org/abs/2404.04808)
- NeuFlow [https://arxiv.org/abs/2403.10425](https://arxiv.org/abs/2403.10425)
- SEA-RAFT [https://arxiv.org/abs/2405.14793](https://arxiv.org/abs/2405.14793)
- SplatFlow [https://arxiv.org/abs/2306.08887](https://arxiv.org/abs/2306.08887)
- Add support for more datasets:
- TartanAir [https://theairlab.org/tartanair-dataset/](https://theairlab.org/tartanair-dataset/)
- Kubric [https://github.com/google-research/kubric](https://github.com/google-research/kubric)
- Add ONNX and TensorRT conversion to RAPIDFlow
- Fix LR scheduler when accumulating gradients
### - v0.3.1
- Added new models:
- CCMR [https://arxiv.org/abs/2311.02661](https://arxiv.org/abs/2311.02661)
- LLA-Flow [https://arxiv.org/abs/2304.08101](https://arxiv.org/abs/2304.08101)
- RAPIDFlow [https://hmorimitsu.com/publication/2024-icra-rapidflow/](https://hmorimitsu.com/publication/2024-icra-rapidflow/)
- Enable FP16 in most models.
- Except the following models, since they have operations that cannot run in FP16: lcv_raft, matchflow, and separableflow
- Add FP16 mode in infer, model_benchmark, and validate scripts
- Create [plot_results.py](plot_results.py) script
- Move resize operations to CUDA (thanks to coca-huang)
### - v0.3.0
- Added new models:
- DIP [https://arxiv.org/abs/2204.00330](https://arxiv.org/abs/2204.00330)
- Flow1D [https://arxiv.org/abs/2103.04524](https://arxiv.org/abs/2103.04524)
- FlowFormer++ [https://arxiv.org/abs/2303.01237](https://arxiv.org/abs/2303.01237)
- GMFlow+, UniMatch [https://arxiv.org/abs/2211.05783](https://arxiv.org/abs/2211.05783)
- MatchFlow [https://arxiv.org/abs/2303.08384](https://arxiv.org/abs/2303.08384)
- MS-RAFT+ [https://arxiv.org/abs/2210.16900](https://arxiv.org/abs/2210.16900)
- RPKNet [https://hmorimitsu.com/publication/2024-aaai-rpknet](https://hmorimitsu.com/publication/2024-aaai-rpknet)
- SeparableFlow [https://openaccess.thecvf.com/content/ICCV2021/papers/Zhang_Separable_Flow_Learning_Motion_Cost_Volumes_for_Optical_Flow_Estimation_ICCV_2021_paper.pdf](https://openaccess.thecvf.com/content/ICCV2021/papers/Zhang_Separable_Flow_Learning_Motion_Cost_Volumes_for_Optical_Flow_Estimation_ICCV_2021_paper.pdf)
- SKFlow [https://arxiv.org/abs/2205.14623](https://arxiv.org/abs/2205.14623)
- VideoFlow [https://arxiv.org/abs/2303.08340](https://arxiv.org/abs/2303.08340)
- `speed_benchmark.py` becomes `model_benchmark.py` and records more metrics
- Fix compatibility with PyTorch 2.0
- Fix compatibility with PyTorch Lightning 1.9
- Fix resizing augmentation when the valid mask is sparse
- Add support for more datasets:
- Middlebury [https://vision.middlebury.edu/flow/](https://vision.middlebury.edu/flow/)
- Monkaa [https://lmb.informatik.uni-freiburg.de/resources/datasets/SceneFlowDatasets.en.html](https://lmb.informatik.uni-freiburg.de/resources/datasets/SceneFlowDatasets.en.html)
- Spring [https://spring-benchmark.org/](https://spring-benchmark.org/)
## Available models
- CCMR [https://arxiv.org/abs/2311.02661](https://arxiv.org/abs/2311.02661)
- CRAFT [https://arxiv.org/abs/2203.16896](https://arxiv.org/abs/2203.16896)
- CSFlow [https://arxiv.org/abs/2202.00909](https://arxiv.org/abs/2202.00909)
- DICL-Flow [https://arxiv.org/abs/2010.14851](https://arxiv.org/abs/2010.14851)
- DIP [https://arxiv.org/abs/2204.00330](https://arxiv.org/abs/2204.00330)
- FastFlowNet [https://arxiv.org/abs/2103.04524](https://arxiv.org/abs/2103.04524)
- Flow1D [https://arxiv.org/abs/2103.04524](https://arxiv.org/abs/2103.04524)
- FlowFormer [https://arxiv.org/abs/2203.16194](https://arxiv.org/abs/2203.16194)
- FlowFormer++ [https://arxiv.org/abs/2303.01237](https://arxiv.org/abs/2303.01237)
- FlowNet [https://arxiv.org/abs/1504.06852](https://arxiv.org/abs/1504.06852)
- FlowNet2 [https://arxiv.org/abs/1612.01925](https://arxiv.org/abs/1612.01925)
- GMA [https://arxiv.org/abs/2104.02409](https://arxiv.org/abs/2104.02409)
- GMFlow [https://arxiv.org/abs/2111.13680](https://arxiv.org/abs/2111.13680)
- GMFlow+, UniMatch [https://arxiv.org/abs/2211.05783](https://arxiv.org/abs/2211.05783)
- GMFlowNet [https://arxiv.org/abs/2203.11335](https://arxiv.org/abs/2203.11335)
- HD3 [https://arxiv.org/abs/1812.06264](https://arxiv.org/abs/1812.06264)
- IRR [https://arxiv.org/abs/1904.05290](https://arxiv.org/abs/1904.05290)
- LCV [https://arxiv.org/abs/2007.11431](https://arxiv.org/abs/2007.11431)
- LiteFlowNet [https://arxiv.org/abs/1805.07036](https://arxiv.org/abs/1805.07036)
- LiteFlowNet2 [https://arxiv.org/abs/1903.07414](https://arxiv.org/abs/1903.07414)
- LiteFlowNet3 [https://arxiv.org/abs/2007.09319](https://arxiv.org/abs/2007.09319)
- LLA-Flow [https://arxiv.org/abs/2304.08101](https://arxiv.org/abs/2304.08101)
- MaskFlownet [https://arxiv.org/abs/2003.10955](https://arxiv.org/abs/2003.10955)
- MatchFlow [https://arxiv.org/abs/2303.08384](https://arxiv.org/abs/2303.08384)
- MemFlow [https://arxiv.org/abs/2404.04808](https://arxiv.org/abs/2404.04808)
- MS-RAFT+ [https://arxiv.org/abs/2210.16900](https://arxiv.org/abs/2210.16900)
- NeuFlow v1 [https://arxiv.org/abs/2403.10425](https://arxiv.org/abs/2403.10425)
- NeuFlow v2 [https://arxiv.org/abs/2408.10161](https://arxiv.org/abs/2408.10161)
- PWCNet [https://arxiv.org/abs/1709.02371](https://arxiv.org/abs/1709.02371)
- RAFT [https://arxiv.org/abs/2003.12039](https://arxiv.org/abs/2003.12039)
- RAPIDFlow [https://hmorimitsu.com/publication/2024-icra-rapidflow/](https://hmorimitsu.com/publication/2024-icra-rapidflow/)
- RPKNet [https://hmorimitsu.com/publication/2024-aaai-rpknet](https://hmorimitsu.com/publication/2024-aaai-rpknet)
- ScopeFlow [https://arxiv.org/abs/2002.10770](https://arxiv.org/abs/2002.10770)
- SCV [https://arxiv.org/abs/2104.02166](https://arxiv.org/abs/2104.02166)
- SEA-RAFT [https://arxiv.org/abs/2405.14793](https://arxiv.org/abs/2405.14793)
- SeparableFlow [https://openaccess.thecvf.com/content/ICCV2021/papers/Zhang_Separable_Flow_Learning_Motion_Cost_Volumes_for_Optical_Flow_Estimation_ICCV_2021_paper.pdf](https://openaccess.thecvf.com/content/ICCV2021/papers/Zhang_Separable_Flow_Learning_Motion_Cost_Volumes_for_Optical_Flow_Estimation_ICCV_2021_paper.pdf)
- SKFlow [https://arxiv.org/abs/2205.14623](https://arxiv.org/abs/2205.14623)
- SplatFlow [https://arxiv.org/abs/2306.08887](https://arxiv.org/abs/2306.08887)
- STaRFlow [https://arxiv.org/abs/2007.05481](https://arxiv.org/abs/2007.05481)
- VCN [https://papers.nips.cc/paper/2019/file/bbf94b34eb32268ada57a3be5062fe7d-Paper.pdf](https://papers.nips.cc/paper/2019/file/bbf94b34eb32268ada57a3be5062fe7d-Paper.pdf)
- VideoFlow [https://arxiv.org/abs/2303.08340](https://arxiv.org/abs/2303.08340)
Read more details about the models on [https://ptlflow.readthedocs.io/en/latest/models/models_list.html](https://ptlflow.readthedocs.io/en/latest/models/models_list.html).
# Results
You can see a table with main evaluation results of the available models [here](https://ptlflow.readthedocs.io/en/latest/results/accuracy_epe.html). More results are also available in the folder [docs/source/results](docs/source/results).
**Disclaimer**: These results are the ones obtained by evaluating the available models in this framework in my machine. Your results may be different due to differences in hardware and software. I also do not guarantee that the results of each model will be similar to the ones presented in the respective papers or other original sources. If you need to replicate the original results from a paper, you should use the original implementations.
## Getting started
Please take a look at the [documentation](https://ptlflow.readthedocs.io/) to learn how to install and use PTLFlow.
You can also check the notebooks below running on Google Colab for some practical examples:
- [Inference with a pretrained model](https://colab.research.google.com/drive/1_WXvIRweQJgex0X-HS0LFXBb0IWZIvR4?usp=sharing).
- [Training and using the learned weights for inference](https://colab.research.google.com/drive/1b_SMGSXh9F9TkinqZt0c64EH-GE87HVi?usp=sharing).
If you are using the previous v0.3.X code, then check the [v0.3.2 documentation](https://ptlflow.readthedocs.io/en/v0.3.2/) and the following example notebooks:
- [Inference with a pretrained model (PTLFlow v0.3)](https://colab.research.google.com/drive/1YARBRUGplqTRnRuY9sKIs6LY_2kWAWZJ?usp=sharing).
- [Training and using the learned weights for inference (PTLFlow v0.3)](https://colab.research.google.com/drive/1mbuAEF728_jZpFEsQHXDxjIGAcB1-nVs?usp=sharing).
## Licenses
The original code of this repository is licensed under the [Apache 2.0 license](LICENSE).
Each model may be subjected to different licenses. The license of each model is included in their respective folders. It is your responsibility to make sure that your project is in compliance with all the licenses and conditions involved.
The external pretrained weights all have different licenses, which are listed in their respective folders.
The pretrained weights that were trained within this project are available under the [CC BY-NC-SA 4.0 license](https://creativecommons.org/licenses/by-nc-sa/4.0/), which I believe that covers the licenses of the datasets used in the training. That being said, I am not a legal expert so if you plan to use them to any purpose other than research, you should check all the involved licenses by yourself. Additionally, the datasets used for the training usually require the user to cite the original papers, so be sure to include their respective references in your work.
## Contributing
Contribution are welcome! Please check [CONTRIBUTING.md](CONTRIBUTING.md) to see how to contribute.
## Citing
### BibTeX
```
@misc{morimitsu2021ptlflow,
author = {Henrique Morimitsu},
title = {PyTorch Lightning Optical Flow},
year = {2021},
publisher = {GitHub},
journal = {GitHub repository},
howpublished = {\url{https://github.com/hmorimitsu/ptlflow}}
}
```
## Acknowledgements
- This README file is heavily inspired by the one from the [timm](https://github.com/rwightman/pytorch-image-models) repository.
- Some parts of the code were inspired by or taken from [FlowNetPytorch](https://github.com/ClementPinard/FlowNetPytorch).
- [flownet2-pytorch](https://github.com/NVIDIA/flownet2-pytorch) was also another important source.
- The current main training routine is based on [RAFT](https://github.com/princeton-vl/RAFT).
Raw data
{
"_id": null,
"home_page": null,
"name": "ptlflow",
"maintainer": null,
"docs_url": null,
"requires_python": "<3.13,>=3.8",
"maintainer_email": null,
"keywords": "torch, pytorch lightning, optical flow, models",
"author": null,
"author_email": "Henrique Morimitsu <henriquem87@gmail.com>",
"download_url": "https://files.pythonhosted.org/packages/6d/aa/72f1f072de776d5c17e1bac169c8cbc1ef06472c87d2378edb3761a9cd27/ptlflow-0.4.0.tar.gz",
"platform": null,
"description": "# PyTorch Lightning Optical Flow\n\n![GitHub CI python status](https://github.com/hmorimitsu/ptlflow/actions/workflows/python.yml/badge.svg)\n![GitHub CI pytorch status](https://github.com/hmorimitsu/ptlflow/actions/workflows/pytorch.yml/badge.svg)\n![GitHub CI lightning status](https://github.com/hmorimitsu/ptlflow/actions/workflows/lightning.yml/badge.svg)\n![GitHub CI build status](https://github.com/hmorimitsu/ptlflow/actions/workflows/build.yml/badge.svg)\n\n## Introduction\n\nThis is a collection of state-of-the-art deep model for estimating optical flow. The main goal is to provide a unified framework where multiple models can be trained and tested more easily.\n\nThe work and code from many others are present here. I tried to make sure everything is properly referenced, but please let me know if I missed something.\n\nThis is still under development, so some things may not work as intended. I plan to add more models in the future, as well keep improving the platform.\n\n- [What's new](#whats-new)\n- [Available models](#available-models)\n- [Results](#results)\n- [Getting started](#getting-started)\n- [Licenses](#licenses)\n- [Contributing](#contributing)\n- [Citing](#citing)\n- [Acknowledgements](#acknowledgements)\n\n## What's new\n\n### - v0.4.0\n\nMajor update to support Lightning 2 (finally!). However, it also introduces breaking changes from the previous v0.3 code. See the details below.\n\n- Transitioning from v0.3 to v0.4: check the [v0.4 upgrade guide](https://ptlflow.readthedocs.io/en/latest/starting/v04_upgrade_guide.html)\n- Added features:\n - Support for YAML config files. See the [config file documentation](https://ptlflow.readthedocs.io/en/latest/starting/config_files.html)\n - Table [comparing PTLFlow results with the original papers](https://ptlflow.readthedocs.io/en/latest/results/paper_ptlflow.html) to check the stability of the included models.\n- Added new models:\n - NeuFlow v2 [https://arxiv.org/abs/2408.10161](https://arxiv.org/abs/2408.10161)\n- Add support for more datasets:\n - Middlebury-ST [https://vision.middlebury.edu/stereo/data/scenes2014/]{https://vision.middlebury.edu/stereo/data/scenes2014/}\n - VIPER [https://playing-for-benchmarks.org/](https://playing-for-benchmarks.org/)\n\n### - v0.3.2\n\n- Added new models:\n - MemFlow [https://arxiv.org/abs/2404.04808](https://arxiv.org/abs/2404.04808)\n - NeuFlow [https://arxiv.org/abs/2403.10425](https://arxiv.org/abs/2403.10425)\n - SEA-RAFT [https://arxiv.org/abs/2405.14793](https://arxiv.org/abs/2405.14793)\n - SplatFlow [https://arxiv.org/abs/2306.08887](https://arxiv.org/abs/2306.08887)\n- Add support for more datasets:\n - TartanAir [https://theairlab.org/tartanair-dataset/](https://theairlab.org/tartanair-dataset/)\n - Kubric [https://github.com/google-research/kubric](https://github.com/google-research/kubric)\n- Add ONNX and TensorRT conversion to RAPIDFlow\n- Fix LR scheduler when accumulating gradients\n\n### - v0.3.1\n\n- Added new models:\n - CCMR [https://arxiv.org/abs/2311.02661](https://arxiv.org/abs/2311.02661)\n - LLA-Flow [https://arxiv.org/abs/2304.08101](https://arxiv.org/abs/2304.08101)\n - RAPIDFlow [https://hmorimitsu.com/publication/2024-icra-rapidflow/](https://hmorimitsu.com/publication/2024-icra-rapidflow/)\n- Enable FP16 in most models.\n - Except the following models, since they have operations that cannot run in FP16: lcv_raft, matchflow, and separableflow\n- Add FP16 mode in infer, model_benchmark, and validate scripts\n- Create [plot_results.py](plot_results.py) script\n- Move resize operations to CUDA (thanks to coca-huang)\n\n### - v0.3.0\n\n- Added new models:\n - DIP [https://arxiv.org/abs/2204.00330](https://arxiv.org/abs/2204.00330)\n - Flow1D [https://arxiv.org/abs/2103.04524](https://arxiv.org/abs/2103.04524)\n - FlowFormer++ [https://arxiv.org/abs/2303.01237](https://arxiv.org/abs/2303.01237)\n - GMFlow+, UniMatch [https://arxiv.org/abs/2211.05783](https://arxiv.org/abs/2211.05783)\n - MatchFlow [https://arxiv.org/abs/2303.08384](https://arxiv.org/abs/2303.08384)\n - MS-RAFT+ [https://arxiv.org/abs/2210.16900](https://arxiv.org/abs/2210.16900)\n - RPKNet [https://hmorimitsu.com/publication/2024-aaai-rpknet](https://hmorimitsu.com/publication/2024-aaai-rpknet)\n - SeparableFlow [https://openaccess.thecvf.com/content/ICCV2021/papers/Zhang_Separable_Flow_Learning_Motion_Cost_Volumes_for_Optical_Flow_Estimation_ICCV_2021_paper.pdf](https://openaccess.thecvf.com/content/ICCV2021/papers/Zhang_Separable_Flow_Learning_Motion_Cost_Volumes_for_Optical_Flow_Estimation_ICCV_2021_paper.pdf)\n - SKFlow [https://arxiv.org/abs/2205.14623](https://arxiv.org/abs/2205.14623)\n - VideoFlow [https://arxiv.org/abs/2303.08340](https://arxiv.org/abs/2303.08340)\n- `speed_benchmark.py` becomes `model_benchmark.py` and records more metrics\n- Fix compatibility with PyTorch 2.0\n- Fix compatibility with PyTorch Lightning 1.9\n- Fix resizing augmentation when the valid mask is sparse\n- Add support for more datasets:\n - Middlebury [https://vision.middlebury.edu/flow/](https://vision.middlebury.edu/flow/)\n - Monkaa [https://lmb.informatik.uni-freiburg.de/resources/datasets/SceneFlowDatasets.en.html](https://lmb.informatik.uni-freiburg.de/resources/datasets/SceneFlowDatasets.en.html)\n - Spring [https://spring-benchmark.org/](https://spring-benchmark.org/)\n\n## Available models\n\n- CCMR [https://arxiv.org/abs/2311.02661](https://arxiv.org/abs/2311.02661)\n- CRAFT [https://arxiv.org/abs/2203.16896](https://arxiv.org/abs/2203.16896)\n- CSFlow [https://arxiv.org/abs/2202.00909](https://arxiv.org/abs/2202.00909)\n- DICL-Flow [https://arxiv.org/abs/2010.14851](https://arxiv.org/abs/2010.14851)\n- DIP [https://arxiv.org/abs/2204.00330](https://arxiv.org/abs/2204.00330)\n- FastFlowNet [https://arxiv.org/abs/2103.04524](https://arxiv.org/abs/2103.04524)\n- Flow1D [https://arxiv.org/abs/2103.04524](https://arxiv.org/abs/2103.04524)\n- FlowFormer [https://arxiv.org/abs/2203.16194](https://arxiv.org/abs/2203.16194)\n- FlowFormer++ [https://arxiv.org/abs/2303.01237](https://arxiv.org/abs/2303.01237)\n- FlowNet [https://arxiv.org/abs/1504.06852](https://arxiv.org/abs/1504.06852)\n- FlowNet2 [https://arxiv.org/abs/1612.01925](https://arxiv.org/abs/1612.01925)\n- GMA [https://arxiv.org/abs/2104.02409](https://arxiv.org/abs/2104.02409)\n- GMFlow [https://arxiv.org/abs/2111.13680](https://arxiv.org/abs/2111.13680)\n- GMFlow+, UniMatch [https://arxiv.org/abs/2211.05783](https://arxiv.org/abs/2211.05783)\n- GMFlowNet [https://arxiv.org/abs/2203.11335](https://arxiv.org/abs/2203.11335)\n- HD3 [https://arxiv.org/abs/1812.06264](https://arxiv.org/abs/1812.06264)\n- IRR [https://arxiv.org/abs/1904.05290](https://arxiv.org/abs/1904.05290)\n- LCV [https://arxiv.org/abs/2007.11431](https://arxiv.org/abs/2007.11431)\n- LiteFlowNet [https://arxiv.org/abs/1805.07036](https://arxiv.org/abs/1805.07036)\n- LiteFlowNet2 [https://arxiv.org/abs/1903.07414](https://arxiv.org/abs/1903.07414)\n- LiteFlowNet3 [https://arxiv.org/abs/2007.09319](https://arxiv.org/abs/2007.09319)\n- LLA-Flow [https://arxiv.org/abs/2304.08101](https://arxiv.org/abs/2304.08101)\n- MaskFlownet [https://arxiv.org/abs/2003.10955](https://arxiv.org/abs/2003.10955)\n- MatchFlow [https://arxiv.org/abs/2303.08384](https://arxiv.org/abs/2303.08384)\n- MemFlow [https://arxiv.org/abs/2404.04808](https://arxiv.org/abs/2404.04808)\n- MS-RAFT+ [https://arxiv.org/abs/2210.16900](https://arxiv.org/abs/2210.16900)\n- NeuFlow v1 [https://arxiv.org/abs/2403.10425](https://arxiv.org/abs/2403.10425)\n- NeuFlow v2 [https://arxiv.org/abs/2408.10161](https://arxiv.org/abs/2408.10161)\n- PWCNet [https://arxiv.org/abs/1709.02371](https://arxiv.org/abs/1709.02371)\n- RAFT [https://arxiv.org/abs/2003.12039](https://arxiv.org/abs/2003.12039)\n- RAPIDFlow [https://hmorimitsu.com/publication/2024-icra-rapidflow/](https://hmorimitsu.com/publication/2024-icra-rapidflow/)\n- RPKNet [https://hmorimitsu.com/publication/2024-aaai-rpknet](https://hmorimitsu.com/publication/2024-aaai-rpknet)\n- ScopeFlow [https://arxiv.org/abs/2002.10770](https://arxiv.org/abs/2002.10770)\n- SCV [https://arxiv.org/abs/2104.02166](https://arxiv.org/abs/2104.02166)\n- SEA-RAFT [https://arxiv.org/abs/2405.14793](https://arxiv.org/abs/2405.14793)\n- SeparableFlow [https://openaccess.thecvf.com/content/ICCV2021/papers/Zhang_Separable_Flow_Learning_Motion_Cost_Volumes_for_Optical_Flow_Estimation_ICCV_2021_paper.pdf](https://openaccess.thecvf.com/content/ICCV2021/papers/Zhang_Separable_Flow_Learning_Motion_Cost_Volumes_for_Optical_Flow_Estimation_ICCV_2021_paper.pdf)\n- SKFlow [https://arxiv.org/abs/2205.14623](https://arxiv.org/abs/2205.14623)\n- SplatFlow [https://arxiv.org/abs/2306.08887](https://arxiv.org/abs/2306.08887)\n- STaRFlow [https://arxiv.org/abs/2007.05481](https://arxiv.org/abs/2007.05481)\n- VCN [https://papers.nips.cc/paper/2019/file/bbf94b34eb32268ada57a3be5062fe7d-Paper.pdf](https://papers.nips.cc/paper/2019/file/bbf94b34eb32268ada57a3be5062fe7d-Paper.pdf)\n- VideoFlow [https://arxiv.org/abs/2303.08340](https://arxiv.org/abs/2303.08340)\n\nRead more details about the models on [https://ptlflow.readthedocs.io/en/latest/models/models_list.html](https://ptlflow.readthedocs.io/en/latest/models/models_list.html).\n\n# Results\n\nYou can see a table with main evaluation results of the available models [here](https://ptlflow.readthedocs.io/en/latest/results/accuracy_epe.html). More results are also available in the folder [docs/source/results](docs/source/results).\n\n**Disclaimer**: These results are the ones obtained by evaluating the available models in this framework in my machine. Your results may be different due to differences in hardware and software. I also do not guarantee that the results of each model will be similar to the ones presented in the respective papers or other original sources. If you need to replicate the original results from a paper, you should use the original implementations.\n\n## Getting started\n\nPlease take a look at the [documentation](https://ptlflow.readthedocs.io/) to learn how to install and use PTLFlow.\n\nYou can also check the notebooks below running on Google Colab for some practical examples:\n\n- [Inference with a pretrained model](https://colab.research.google.com/drive/1_WXvIRweQJgex0X-HS0LFXBb0IWZIvR4?usp=sharing).\n- [Training and using the learned weights for inference](https://colab.research.google.com/drive/1b_SMGSXh9F9TkinqZt0c64EH-GE87HVi?usp=sharing).\n\nIf you are using the previous v0.3.X code, then check the [v0.3.2 documentation](https://ptlflow.readthedocs.io/en/v0.3.2/) and the following example notebooks:\n\n- [Inference with a pretrained model (PTLFlow v0.3)](https://colab.research.google.com/drive/1YARBRUGplqTRnRuY9sKIs6LY_2kWAWZJ?usp=sharing).\n- [Training and using the learned weights for inference (PTLFlow v0.3)](https://colab.research.google.com/drive/1mbuAEF728_jZpFEsQHXDxjIGAcB1-nVs?usp=sharing).\n\n## Licenses\n\nThe original code of this repository is licensed under the [Apache 2.0 license](LICENSE).\n\nEach model may be subjected to different licenses. The license of each model is included in their respective folders. It is your responsibility to make sure that your project is in compliance with all the licenses and conditions involved.\n\nThe external pretrained weights all have different licenses, which are listed in their respective folders.\n\nThe pretrained weights that were trained within this project are available under the [CC BY-NC-SA 4.0 license](https://creativecommons.org/licenses/by-nc-sa/4.0/), which I believe that covers the licenses of the datasets used in the training. That being said, I am not a legal expert so if you plan to use them to any purpose other than research, you should check all the involved licenses by yourself. Additionally, the datasets used for the training usually require the user to cite the original papers, so be sure to include their respective references in your work.\n\n## Contributing\n\nContribution are welcome! Please check [CONTRIBUTING.md](CONTRIBUTING.md) to see how to contribute.\n\n## Citing\n\n### BibTeX\n\n```\n@misc{morimitsu2021ptlflow,\n author = {Henrique Morimitsu},\n title = {PyTorch Lightning Optical Flow},\n year = {2021},\n publisher = {GitHub},\n journal = {GitHub repository},\n howpublished = {\\url{https://github.com/hmorimitsu/ptlflow}}\n}\n```\n\n## Acknowledgements\n\n- This README file is heavily inspired by the one from the [timm](https://github.com/rwightman/pytorch-image-models) repository.\n- Some parts of the code were inspired by or taken from [FlowNetPytorch](https://github.com/ClementPinard/FlowNetPytorch).\n- [flownet2-pytorch](https://github.com/NVIDIA/flownet2-pytorch) was also another important source.\n- The current main training routine is based on [RAFT](https://github.com/princeton-vl/RAFT).\n",
"bugtrack_url": null,
"license": "Apache 2.0 License",
"summary": "PyTorch Lightning Optical Flow",
"version": "0.4.0",
"project_urls": {
"Documentation": "https://ptlflow.readthedocs.io/",
"Source Code": "https://github.com/hmorimitsu/ptlflow"
},
"split_keywords": [
"torch",
" pytorch lightning",
" optical flow",
" models"
],
"urls": [
{
"comment_text": "",
"digests": {
"blake2b_256": "bcf75da218d3a0f0e4bff5a09da610f071b6d14553c7bbfcf10e787dca2bf11f",
"md5": "2135f5c224d28385dc22778dfffb4fa5",
"sha256": "90282686667a860d097d4be809d0fed8842f9b57431e8646db5a871d8c3eb949"
},
"downloads": -1,
"filename": "ptlflow-0.4.0-py3-none-any.whl",
"has_sig": false,
"md5_digest": "2135f5c224d28385dc22778dfffb4fa5",
"packagetype": "bdist_wheel",
"python_version": "py3",
"requires_python": "<3.13,>=3.8",
"size": 898926,
"upload_time": "2024-12-04T03:26:01",
"upload_time_iso_8601": "2024-12-04T03:26:01.283451Z",
"url": "https://files.pythonhosted.org/packages/bc/f7/5da218d3a0f0e4bff5a09da610f071b6d14553c7bbfcf10e787dca2bf11f/ptlflow-0.4.0-py3-none-any.whl",
"yanked": false,
"yanked_reason": null
},
{
"comment_text": "",
"digests": {
"blake2b_256": "6daa72f1f072de776d5c17e1bac169c8cbc1ef06472c87d2378edb3761a9cd27",
"md5": "bf0a39295276c2a3e93747efb3ebc85c",
"sha256": "99c6db4da09528cce72d18eba4efcc0dbcd96a88bddf209b8ca00eac69e32022"
},
"downloads": -1,
"filename": "ptlflow-0.4.0.tar.gz",
"has_sig": false,
"md5_digest": "bf0a39295276c2a3e93747efb3ebc85c",
"packagetype": "sdist",
"python_version": "source",
"requires_python": "<3.13,>=3.8",
"size": 722873,
"upload_time": "2024-12-04T03:26:02",
"upload_time_iso_8601": "2024-12-04T03:26:02.712761Z",
"url": "https://files.pythonhosted.org/packages/6d/aa/72f1f072de776d5c17e1bac169c8cbc1ef06472c87d2378edb3761a9cd27/ptlflow-0.4.0.tar.gz",
"yanked": false,
"yanked_reason": null
}
],
"upload_time": "2024-12-04 03:26:02",
"github": true,
"gitlab": false,
"bitbucket": false,
"codeberg": false,
"github_user": "hmorimitsu",
"github_project": "ptlflow",
"travis_ci": false,
"coveralls": false,
"github_actions": true,
"requirements": [
{
"name": "einops",
"specs": [
[
"<",
"0.9"
]
]
},
{
"name": "h5py",
"specs": [
[
"<",
"3.13"
]
]
},
{
"name": "kaleido",
"specs": [
[
"<",
"0.3"
]
]
},
{
"name": "lightning",
"specs": [
[
">=",
"2.1"
],
[
"<",
"2.5"
]
]
},
{
"name": "loguru",
"specs": [
[
"<",
"0.8"
]
]
},
{
"name": "opencv-python",
"specs": [
[
"<",
"4.11"
]
]
},
{
"name": "pandas",
"specs": [
[
"<",
"2.3"
]
]
},
{
"name": "plotly",
"specs": [
[
"<",
"5.25"
]
]
},
{
"name": "pypng",
"specs": [
[
"==",
"0.20220715.0"
]
]
},
{
"name": "scipy",
"specs": [
[
"<",
"1.15"
]
]
},
{
"name": "tabulate",
"specs": [
[
"<",
"0.10"
]
]
},
{
"name": "tensorboard",
"specs": [
[
"<",
"2.19"
]
]
},
{
"name": "timm",
"specs": [
[
"<",
"1.1"
]
]
}
],
"lcname": "ptlflow"
}