Cho-Ying Wu, Jialiang Wang, Michael Hall, Ulrich Neumann, Shuochen Su
Our DistDepth is a highly robust monocular depth estimation approach for generic indoor scenes.
- Trained with stereo sequences without their groundtruth depth
- Structured and metric-accurate
- Run in an interactive rate with Laptop GPU
- Sim-to-real: trained on simulation and becomes transferrable to real scenes
We test on Ubuntu 20.04 LTS with an NVIDIA 2080 (only GPU is supported).
Install packages
-
Use conda
conda create --name distdepth python=3.8
conda activate distdepth
-
Install pre-requisite common packages. Go to https://pytorch.org/get-started/locally/ and install pytorch that is compatible to your computer. We test on pytorch v1.9.0 and cudatoolkit-11.1.
conda install pytorch==1.9.0 torchvision==0.10.0 torchaudio==0.9.0 cudatoolkit=11.3 -c pytorch -c conda-forge
-
Install other dependencies: opencv-python and matplotlib.
pip install opencv-python, matplotlib
Download pretrained models
-
Download pretrained models [here] (ResNet152, 246MB).
-
Move the downloaded item under this folder, and then unzip it. You should be able to see a new folder 'ckpts' that contains the pretrained models.
-
Run
python demo.py
-
Results will be stored under
results/
Download SimSIN [here]
@inproceedings{wu2022toward,
title={Toward Practical Monocular Indoor Depth Estimation},
author={Wu, Cho-Ying and Wang, Jialiang and Hall, Michael and Neumann, Ulrich and Su, Shuochen},
booktitle={CVPR},
year={2022}
}
DistDepth is CC-BY-NC licensed, as found in the LICENSE file.