This repository contains the official PyTorch implementation of the paper "Learning by Analogy: Reliable Supervision from Transformations for Unsupervised Optical Flow Estimation".
For any inquiries, please contact Liang Liu at [email protected]
This code has been developed under Python3, PyTorch 1.1.0 and CUDA 9.0 on Ubuntu 16.04.
We strongly recommend that using docker to ensure you can get the same results as us. The Dockerfile is available. Also, you can build the environment by following:
# Install python packages
pip3 install -r requirements.txt
# Compile the coorelation package with gcc and g++ >= 4.9
cd ./models/correlation_package
python3 setup.py install
# Additional dependencies for training
sudo apt-get update && apt-get install -y libsm6 libxext6 libxrender-dev
pip3 install 'opencv-python>=3.0,<4.0' path.py tensorboardX fast_slic
If you have any trouble with the correlation package, we also provide an alternative implementation. You can modify the import lines in models/pwclite.py to use it.
The checkpoints folder contains the pre-trained models of ARFlow and ARFlow-mv for various datasets.
A minimal example of using a pre-trained is given in inference.py. For two-view models, just run with:
python3 inference.py -m checkpoints/KITTI15/pwclite_ar.tar -s 384 640 \
-i examples/img1.png examples/img2.png
For multi-view model, input with 3 frames:
python3 inference.py -m checkpoints/KITTI15/pwclite_ar_mv.tar -s 384 640 \
-i examples/img0.png examples/img1.png examples/img2.png
We recommend input with 384x640 for KITTI and Cityscapes models, 448x1024 for Sintel models.
Here we provide the complete training pipeline for ARFlow on Sintel and KITTI datasets:
-
Pre-train on the Sintel raw movie. Also, you can skip this step with our pretrained model.
python3 train.py -c sintel_raw.json
-
Fine-tune on the Sintel training set. We provide both settings for training with or without AR for your convenience.
# without AR python3 train.py -c sintel_ft.json # with AR python3 train.py -c sintel_ft_ar.json
The default configuration uses the whole training set for training and validation. We strongly recommend you use the re-split sets as in our ablation studies. You can modify the config file by setting
train_subsplit
to 'train' and settingval_subsplit
to 'val'.
The pipeline is similar to Sintel, refer to configs for more details.
You can pre-train on KITTI raw data, and then fine-tuning on the multi-view extension, or directly train on the multi-view extension. The final results should be similar.
Also, a complete evaluation for a model can be simply run with the option -e
, for example:
python3 train.py -c configs/sintel_ft.json -m checkpoints/Sintel/pwclite_ar.tar -e
Due to copyright issues, please download the dataset from the official websites.
We have upload the Sintel Raw dataset to Google Drive and Baidu Cloud (Key: mxe9). This dataset is created by manually dividing all frames into folders by shot. It should follow the license in the Sintel benchmark.
If you think this work is useful for your research, please consider citing:
@inproceedings{liu2020learning,
title = {Learning by Analogy: Reliable Supervision from Transformations for Unsupervised Optical Flow Estimation},
author = {Liu, Liang and Zhang, Jiangning and He, Ruifei and Liu, Yong and Wang, Yabiao and Tai, Ying and Luo, Donghao and Wang, Chengjie and Li, Jilin and Huang, Feiyue},
booktitle = {IEEE Conference on Computer Vision and Pattern Recognition(CVPR)},
year = {2020}
}
We thank Pengpeng Liu for in-depth discussions and helpful comments. Also, we thank for portions of the source code from some great works such as Fast-SLIC, IRR.