Skip to content

Latest commit

 

History

History
61 lines (49 loc) · 2.33 KB

README.md

File metadata and controls

61 lines (49 loc) · 2.33 KB

MIMDepth

This is the official code for ICRA 2023 paper Image Masking for Robust Self-Supervised Monocular Depth Estimation by Hemang Chawla, Kishaan Jeeveswaran, Elahe Arani and Bahram Zonooz.

alt text

We propose MIMDepth, a method that adapts masked image modeling (MIM) for self-supervised monocular depth estimation

Install

MIMDepth was trained on TeslaV100 GPU for 20 epochs with AdamW optimizer at a resolution of (192 x 640) with a batchsize 12. The docker environment used can be setup as follows:

git clone https://github.com/NeurAI-Lab/MIMDepth.git
cd MIMDepth
make docker-build

Training

MIMDepth is trained in a self-supervised manner from videos. For training, utilize a .yaml config file or a .ckpt model checkpoint file with scripts/train.py.

python scripts/train.py <config_file.yaml or model_checkpoint.ckpt>

Example config file to train MIMDepth can be found in configs folder.

Evaluation

A trained model can be evaluated by providing a .ckpt model checkpoint.

python scripts/eval.py --checkpoint <model_checkpoint.ckpt>

For running inference on a single image or folder,

python scripts/infer.py --checkpoint <checkpoint.ckpt> --input <image or folder> --output <image or folder> [--image_shape <input shape (h,w)>]

Pretrained Models for MT-SfMLearner and MIMDepth can be found here.

Cite Our Work

If you find the code useful in your research, please consider citing our paper:

@inproceedings{chawla2023image,
	author={H. {Chawla} and K. {Jeeveswaran} and E. {Arani} and B. {Zonooz}},
	booktitle={2023 IEEE International Conference on Robotics and Automation (ICRA)},
	title={Image Masking for Robust Self-Supervised Monocular Depth Estimation},
	location={London, UK},
	publisher={IEEE},
	year={2023}
}

License

This project is licensed under the terms of the MIT license.