• Stars
    star
    111
  • Rank 314,510 (Top 7 %)
  • Language
    Jupyter Notebook
  • License
    MIT License
  • Created almost 5 years ago
  • Updated almost 4 years ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

Deep learning model to estimate the depth of image.

Monocular Depth Estimation with Transfer Learning pretrained MobileNetV2

This project implements a deep learning neural network model to generate the depth image of a given image. Model is a U-net model with MobileNetV2 as the encoder, and model has utilized skip connection from encoder to decoder. Model generates a depth image of resolution 480x640 for input image of same size.

Results

This project was implemented taking reference from the following paper:

High Quality Monocular Depth Estimation via Transfer Learning (arXiv 2018) [Ibraheem Alhashim] and Peter Wonka

Getting Started

Model is trained using the IPYTHON file "train_mobilenetv2.ipynb".
  • Download the dataset and give the location of dataset.
  • Change the following according to the needs: batch_size, epochs, lr (learning rate). Load the pretrained model if needed.
IPYTHON file "test_img.ipynb" can be used to generate the depth image on pretrained model.
  • Give the location for the dictionary of images to be converted and load the pretrained model
IPYTHON file "test_video.ipynb" can be used to generate the depth video on pretrained model.
  • Give the location for the dictionary of images to be converted and load the pretrained model.

Implementation of the Depth estimation using Densenet model is in the folder "Densenet_depth_model".

Dataset

  • NYU Depth V2 (50K) (4.1 GB): File is extraced while running the "train_mobilenetv2.ipynb".

Download the pretrained model

  • Mobilenet (55 MB). Pretrained model is trained on 2 NVIDIA GeForce GTX 1080 for 6 hours(6 epoches).

Author

Written by Alinstein Jose, University of Victoria.