• Stars
    star
    123
  • Rank 290,145 (Top 6 %)
  • Language
    Python
  • License
    Other
  • Created over 2 years ago
  • Updated over 2 years ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

P3Depth: Monocular Depth Estimation with a Piecewise Planarity Prior

Vaishakh Patil, Dr. Christos Sakaridis, Dr. Alex Liniger and Prof. Luc Van Gool

CVPR 2022 (pdf) | arXiv (pdf) | Project page

This repository is still being updated !!!

This is the reference PyTorch implementation for training and evaluation of depth estimation models using the method described in our paper.

example input output

License

This software is released under a creative commons license which allows for personal and research use only. For a commercial license please contact the authors. You can view a license summary here.

P3Depth: Monocular Depth Estimation with a Piecewise Planarity Prior

example input output

Our end-to-end P3Depth method uses the offset vector field to define interactions between pixels within a plane coefficient representation. The plane coefficients of seed pixels are used to predict depth at each position. The resulting prediction is adaptively fused with the initial prediction based on the confidence map to account for potential deviations from precise local planarity.

Contents

  1. Installation
  2. Training
  3. Evaluation
  4. Citation
  5. Contributions
  6. Acknowledgement

Installation

For setup, you need:

  1. Linux
  2. NVIDIA GPU with CUDA & CuDNN
  3. Python 3
  4. Conda

Use the package manager pip to install foobar.

conda create -n p3depth python=3.7
conda activate p3depth
conda install pytorch==1.7.0 torchvision==0.8.0 torchaudio==0.7.0 cudatoolkit=9.2 -c pytorch
pip install -r requirements.txt

Dataset

For data preparation of NYU Depth v2 and KITTI datasets, we follow "Lee et al., From Big to Small: Multi-Scale Local Planar Guidance for Monocular Depth Estimation, arXiv 2019". Refer to From Big to Small git repository for more information.

  • We follow From Big to Small for train/test split for both NYU Depth v2 and KITTI datasets.
  • The train/test list files should be named train.txt and test.txt respectively.
  • The dataset path in .yaml files located in config/dataset directory should point to a directory containing train.txt and test.txt.

Training

  • Before staring training, check config files for correct parameters. The path to output directory is located in config/default.yaml. The default parameters for model are located in config/model directory and for dataset are in config/dataset directory.
    • For training, use experiments/train.sh file. Set correct MODEL_CONFIG, DATASET_CONFIG, and EXP_CONFIG files. You may have to change this file as per your systems training protocol along with L113 trainer.py and "check_machine" method in src/utils.py
  • The experiments can be launched with.
cd experiments
./train.sh
  • More details on the training procedure will be released soon.

Evaluation

  • For evaluation, use the experiments/train.sh by setting "--test" flag as follows:
python3 -u trainer.py --test --model_config ${MODEL_CONFIG} --dataset_config ${DATASET_CONFIG} --exp_config ${EXP_CONFIG}
  • Use the standard test split for NYU Depth v2 and KITTI as described in Dataset section.
  • The evaluation can be launched with.
cd experiments
./train.sh
  • More details on evaluation and pretrained models will be released soon.

Citation

If you find our work useful in your research please consider citing our publication:

@inproceedings{P3Depth,
  author    = {Patil, Vaishakh and Sakaridis, Christos and Liniger, Alex and Van Gool, Luc},
  title     = {P3Depth: Monocular Depth Estimation with a Piecewise Planarity Prior},
  booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
  year      = {2022},
}

Contributions

If you find any bug in the code. Please report to
Vaishakh Patil (patil_at_vision.ee.ethz.ch)

Acknowledgement

This work was supported by Toyota through project TRACE Zurich (Toyota Research on Automated Cars in Europe - Zurich). We thank authors following repositories for sharing the code: From Big to Small, Structure-Guided Ranking Loss, Virtual Normal Loss, Revisiting Single Image Depth Estimation.

More Repositories

1

sam-hq

Segment Anything in High Quality [NeurIPS 2023]
Python
3,689
star
2

sam-pt

SAM-PT: Extending SAM to zero-shot video segmentation with point-based tracking.
Python
970
star
3

transfiner

Mask Transfiner for High-Quality Instance Segmentation, CVPR 2022
Python
525
star
4

qd-3dt

Official implementation of Monocular Quasi-Dense 3D Object Tracking, TPAMI 2022
Python
515
star
5

qdtrack

Quasi-Dense Similarity Learning for Multiple Object Tracking, CVPR 2021 (Oral)
Python
382
star
6

pcan

Prototypical Cross-Attention Networks for Multiple Object Tracking and Segmentation, NeurIPS 2021 Spotlight
Python
362
star
7

MaskFreeVIS

Mask-Free Video Instance Segmentation [CVPR 2023]
Python
358
star
8

bdd100k-models

Model Zoo of BDD100K Dataset
Python
285
star
9

idisc

iDisc: Internal Discretization for Monocular Depth Estimation [CVPR 2023]
Python
279
star
10

LiDAR_snow_sim

LiDAR snowfall simulation
Python
172
star
11

r3d3

Python
144
star
12

shift-dev

SHIFT Dataset DevKit - CVPR2022
Python
103
star
13

cascade-detr

[ICCV'23] Cascade-DETR: Delving into High-Quality Universal Object Detection
Python
92
star
14

tet

Implementation of Tracking Every Thing in the Wild, ECCV 2022
Python
69
star
15

TrafficBots

TrafficBots: Towards World Models for Autonomous Driving Simulation and Motion Prediction. ICRA 2023. Code is now available at https://github.com/zhejz/TrafficBots
51
star
16

nutsh

A Platform for Visual Learning from Human Feedback
TypeScript
42
star
17

vmt

Video Mask Transfiner for High-Quality Video Instance Segmentation (ECCV'2022)
Jupyter Notebook
29
star
18

spc2

Instance-Aware Predictive Navigation in Multi-Agent Environments, ICRA 2021
Python
20
star
19

CISS

Unsupervised condition-level adaptation for semantic segmentation
Python
20
star
20

shift-detection-tta

This repository implements continuous test-time adaptation algorithms for object detection on the SHIFT dataset.
Python
18
star
21

vis4d

A modular library for visual 4D scene understanding
Python
17
star
22

dla-afa

Official implementation of Dense Prediction with Attentive Feature Aggregation, WACV 2023
Python
12
star
23

soccer-player

Python
8
star
24

project-template

Python
4
star
25

vis4d_cuda_ops

Cuda
3
star
26

vis4d-template

Vis4D Template.
Shell
3
star