SqueezeDet: Unified, Small, Low Power Fully Convolutional Neural Networks for Real-Time Object Detection for Autonomous Driving
By Bichen Wu, Alvin Wan, Forrest Iandola, Peter H. Jin, Kurt Keutzer (UC Berkeley & DeepScale)
This repository contains a tensorflow implementation of SqueezeDet, a convolutional neural network based object detector described in our paper: https://arxiv.org/abs/1612.01051. If you find this work useful for your research, please consider citing:
@inproceedings{squeezedet,
Author = {Bichen Wu and Forrest Iandola and Peter H. Jin and Kurt Keutzer},
Title = {SqueezeDet: Unified, Small, Low Power Fully Convolutional Neural Networks for Real-Time Object Detection for Autonomous Driving},
Journal = {arXiv:1612.01051},
Year = {2016}
}
Installation:
The following instructions are written for Linux-based distros.
-
Clone the SqueezeDet repository:
git clone https://github.com/BichenWuUCB/squeezeDet.git
Let's call the top level directory of SqueezeDet
$SQDT_ROOT
. -
(Optional) Setup your own virtual environment.
- The following assumes
python
is the Python2.7 executable. Navigate to your user home directory, and create the virtual environment there.
cd ~ virtualenv env --python=python
- Launch the virtual environment.
source env/bin/activate
- The following assumes
-
Use pip to install required Python packages:
pip install -r requirements.txt
Demo:
-
Download SqueezeDet model parameters from here, untar it, and put it under
$SQDT_ROOT/data/
If you are using command line, type:cd $SQDT_ROOT/data/ wget https://www.dropbox.com/s/a6t3er8f03gdl4z/model_checkpoints.tgz tar -xzvf model_checkpoints.tgz rm model_checkpoints.tgz
-
Now we can run the demo. To detect the sample image
$SQDT_ROOT/data/sample.png
,cd $SQDT_ROOT/ python ./src/demo.py
If the installation is correct, the detector should generate this image:
To detect other image(s), use the flag
--input_path=./data/*.png
to point to input image(s). Input image(s) will be scaled to the resolution of 1242x375 (KITTI image resolution), so it works best when original resolution is close to that. -
SqueezeDet is a real-time object detector, which can be used to detect videos. The video demo will be released later.
Training/Validation:
-
Download KITTI object detection dataset: images and labels. Put them under
$SQDT_ROOT/data/KITTI/
. Unzip them, then you will get two directories:$SQDT_ROOT/data/KITTI/training/
and$SQDT_ROOT/data/KITTI/testing/
. -
Now we need to split the training data into a training set and a vlidation set.
cd $SQDT_ROOT/data/KITTI/ mkdir ImageSets cd ./ImageSets ls ../training/image_2/ | grep ".png" | sed s/.png// > trainval.txt
trainval.txt
contains indices to all the images in the training data. In our experiments, we randomly split half of indices intrainval.txt
intotrain.txt
to form a training set and rest of them intoval.txt
to form a validation set. For your convenience, we provide a script to split the train-val set automatically. Simply runcd $SQDT_ROOT/data/ python random_split_train_val.py
then you should get the
train.txt
andval.txt
under$SQDT_ROOT/data/KITTI/ImageSets
.When above two steps are finished, the structure of
$SQDT_ROOT/data/KITTI/
should at least contain:$SQDT_ROOT/data/KITTI/ |->training/ | |-> image_2/00****.png | L-> label_2/00****.txt |->testing/ | L-> image_2/00****.png L->ImageSets/ |-> trainval.txt |-> train.txt L-> val.txt
-
Next, download the CNN model pretrained for ImageNet classification:
cd $SQDT_ROOT/data/ # SqueezeNet wget https://www.dropbox.com/s/fzvtkc42hu3xw47/SqueezeNet.tgz tar -xzvf SqueezeNet.tgz # ResNet50 wget https://www.dropbox.com/s/p65lktictdq011t/ResNet.tgz tar -xzvf ResNet.tgz # VGG16 wget https://www.dropbox.com/s/zxd72nj012lzrlf/VGG16.tgz tar -xzvf VGG16.tgz
-
Now we can start training. Training script can be found in
$SQDT_ROOT/scripts/train.sh
, which contains commands to train 4 models: SqueezeDet, SqueezeDet+, VGG16+ConvDet, ResNet50+ConvDet.cd $SQDT_ROOT/ ./scripts/train.sh -net (squeezeDet|squeezeDet+|vgg16|resnet50) -train_dir /tmp/bichen/logs/squeezedet -gpu 0
Training logs are saved to the directory specified by
-train_dir
. GPU id is specified by-gpu
. Network to train is specificed by-net
-
Before evaluation, you need to first compile the official evaluation script of KITTI dataset
cd $SQDT_ROOT/src/dataset/kitti-eval make
-
Then, you can launch the evaluation script (in parallel with training) by
cd $SQDT_ROOT/ ./scripts/eval.sh -net (squeezeDet|squeezeDet+|vgg16|resnet50) -eval_dir /tmp/bichen/logs/squeezedet -image_set (train|val) -gpu 1
Note that
-train_dir
in the training script should be the same as-eval_dir
in the evaluation script to make it easy for tensorboard to load logs.You can run two evaluation scripts to simultaneously evaluate the model on training and validation set. The training script keeps dumping checkpoint (model parameters) to the training directory once every 1000 steps (step size can be changed). Once a new checkpoint is saved, evaluation threads load the new checkpoint file and evaluate them on training and validation set.
-
Finally, to monitor training and evaluation process, you can use tensorboard by
tensorboard --logdir=$LOG_DIR
Here,
$LOG_DIR
is the directory where your training and evaluation threads dump log events, which should be the same as-train_dir
and-eval_dir
specified intrain.sh
andeval.sh
. From tensorboard, you should be able to see a lot of information including loss, average precision, error analysis, example detections, model visualization, etc.