Monocular-Depth-Estimation-Toolbox
Introduction
Monocular-Depth-Estimation-Toolbox is an open source monocular depth estimation toolbox based on PyTorch and MMSegmentation v0.16.0.
It aims to benchmark MonoDepth methods and provides effective supports for evaluating and visualizing results.
Major features
-
Unified benchmark
Provide a unified benchmark toolbox for various depth estimation methods.
-
Modular design
Depth estimation frameworks are decomposed into different components. One can easily construct a customized framework by combining different modules.
-
Support of multiple methods out of box
I would like to reproduce some of the most excellent depth estimation methods based on this toolbox.
-
High efficiency
It seems that there are few depth estimation benchmarks, so I start this project and hope it is helpful for research.
Thanks to MMSeg, we own these major features.
Benchmark and model zoo
Results and models are available in the model zoo.
Supported backbones (partially release):
- ResNet (CVPR'2016)
- EfficientNet (ICML'2019)
- Vision Transformer (ICLR'2021)
- Swin Transformer (ICCV'2021)
- I recommend cross-package import in config, so that you can utilize other backbone in MMcls, MMseg, etc. Refer to introduction. I will add more backbones in the future.
Supported methods:
- BTS (Arxiv'2019)
- Adabins (CVPR'2021)
- SimIPU (AAAI'2022)
- DPT (ICCV'2021) (TBD)
- DepthFormer (Arxiv'2022)
Supported datasets:
Installation
Please refer to get_started.md for installation and dataset_prepare.md for dataset preparation.
Get Started
We provide train.md and inference.md for the usage of this toolbox.
License
This project is released under the Apache 2.0 license.
Acknowledgement
This repo benefits from awesome works of mmsegmentation, Adabins, BTS. Please also consider citing them.
Cite
@article{li2021simipu,
title={SimIPU: Simple 2D Image and 3D Point Cloud Unsupervised Pre-Training for Spatial-Aware Visual Representations},
author={Li, Zhenyu and Chen, Zehui and Li, Ang and Fang, Liangji and Jiang, Qinhong and Liu, Xianming and Jiang, Junjun and Zhou, Bolei and Zhao, Hang},
journal={arXiv preprint arXiv:2112.04680},
year={2021}
}
Changelog
- Apr 16: Finish most of docs and provide all trained parameters. Release codes about BTS, Adabins, DPT, SimIPU, and DepthFormer. Support KITTI, NYU-v2, SUN RGB-D(eval), and CityScapes.
TODO
- Some annotations in codes are futile, waiting to be rewritten.
- I will release codes of BinsFormer soon.
- I would like to include self-supervised depth estimation methods, such as MonoDepth2.