This code implements the following CVPR 2016 paper (accepted for oral presentation):
Title: "LocNet: Improving Localization Accuracy for Object Detection"
Authors: Spyros Gidaris, Nikos Komodakis
Institution: Universite Paris Est, Ecole des Ponts ParisTech
ArXiv Link: http://arxiv.org/abs/1511.07763
Code: https://github.com/gidariss/LocNet
Abstract:
"We propose a novel object localization methodology with the purpose of boosting the localization accuracy of state-of-the-art object detection systems. Our model, given a search region, aims at returning the bounding box of an object of interest inside this region. To accomplish its goal, it relies on assigning conditional probabilities to each row and column of this region, where these probabilities provide useful information regarding the location of the boundaries of the object inside the search region and allow the accurate inference of the object bounding box under a simple probabilistic framework.
For implementing our localization model, we make use of a convolutional neural network architecture that is properly adapted for this task, called LocNet. We show experimentally that LocNet achieves a very significant improvement on the mAP for high IoU thresholds on PASCAL VOC2007 test set and that it can be very easily coupled with recent state-of-the-art object detection systems, helping them to boost their performance. Finally, we demonstrate that our detection approach can achieve high detection accuracy even when it is given as input a set of sliding windows, thus proving that it is independent of region proposal methods."
If you find LocNet useful in your research, please consider citing:
@inproceedings{gidaris2016locnet,
title={LocNet: Improving Localization Accuracy for Object Detection},
author={Gidaris, Spyros and Komodakis, Nikos},
booktitle={Computer Vision and Pattern Recognition (CVPR), 2016 IEEE Conference on},
year={2016}
}
This code is released under the MIT License (refer to the LICENSE file for details).
Hardware:
For training the LocNet models or testing the LocNet object detection pipeline you will require a GPU with at least 6 Gbytes of memory.
Software:
- Modified version of Caffe developed to supprot LocNet and installed with the cuDNN library [link].
- MATLAB (tested with R2014b)
Optional:
The following packages are necessary for using the EdgeBox or Selective Search bounding box proposas algorithms:
- Edge Boxes code [link].
- The image processing MATLAB toolbox of Piotr Dollar (used for the Edge Boxes) [link].
- Selective search code [link].
Note: we provide the bounding box proposals of PASCAL images and hence installing the above packages is not necessary for training or testing models on this dataset. However, they are necessary for running the demo code on images other than the one that is provided by us.
- Download and install this modified version of Caffe developed to supprot LocNet. To clone Caffe on your local machine:
# $caffe_LocNet: directory where Caffe will be cloned git clone https://github.com/gidariss/caffe_LocNet $caffe_LocNet
$caffe_LocNet
is the directory where Caffe is cloned. After cloning Caffe follow the installation instructions here. Note that you have to install Caffe with the cuDNN library.
2. Clone the LocNet code in your local machine:
```Shell
# $LocNet: directory where LocNet will be cloned
git clone https://github.com/gidariss/LocNet $LocNet
```
From now on, the directory where LocNet was cloned will be called $LocNet
.
3. Create a symbolic link of Caffe installatation directory at $LocNet/external/
:
```Shell
# $LocNet: directory where LocNet was cloned
# $caffe_LoNet: directory where caffe was cloned
ln -sf $caffe_LoNet $LocNet/external/caffe_LocNet
```
- open matlab from the
$LocNet/
directory:cd $LocNet matlab
- Run the
LocNet_build.m
script on matlab command line# matlab command line enviroment >> LocNet_build
Do not worry about the warning messages. They also appear on my machine.
Download the tar.gz files with the pre-trained models:
Recognition models:
- Reduced MR-CNN recognition model.
- Fast RCNN recognition model (re-implemented by us).
Localization models: - LocNet In-Out model.
- LocNet Borders model.
- LocNet Combined model.
- CNN-based bounding box regression model.
Untar and unzip all of the above files on the following locations:
# Recognition models:
$LocNet/models-exps/VGG16_Reduced_MRCNN # Reduced MR-CNN recognition model
$LocNet/models-exps/VGG16_FastRCNN # Fast RCNN recognition model
# VOC2012 structure:
$LocNet/models-exps/VGG16_LocNet_InOut # LocNet In-Out model
$LocNet/models-exps/VGG16_LocNet_Borders # LocNet Borders model
$LocNet/models-exps/VGG16_LocNet_Combined # LocNet Combined model
$LocNet/models-exps/VGG16_BBoxReg # CNN-based bounding box regression model
All of the above models are based on the VGG16-Net and are trained on the union of VOC2007 train+val plus VOC2012 train+val datasets. Note that they are not the same as those used to report result in the paper and hense their performance is slightly different from them (around 0.2 mAP points difference more or less in the percentage scale).
After having complete the basic installation, you will be able to run the demo of object detection based on the LocNet localization models. In order to do that, open the matlab from $LocNet/
directory and then run the script 'demo_LocNet_object_detection_pipeline'
from the matlab command line enviroment:
cd $LocNet
matlab
# matlab command line enviroment
>> demo_LocNet_object_detection_pipeline
Note: you will require a GPU with at least 6 Gbytes of memory in order to run the demo.
In order to play with the parameters of the object detection pipeline read and edit the demo script:
# matlab command line enviroment
>> edit demo_LocNet_object_detection_pipeline.m
First install 1) Edge Boxes, 2) Dollar's image processing MATLAB toolbox (used in Edge Boxes), and 3) Selective Search. Then, create symbolic links of the installation directories at $LocNet/external/
:
# $edges: installation directory of Edge Boxes
ln -sf $edges $LocNet/external/edges
# $pdollar-toolbox: installation directory of Dollar's image processing MATLAB toolbox
ln -sf $pdollar-toolbox $LocNet/external/pdollar-toolbox
# $selective_search: installation directory of selective search code
ln -sf $selective_search $LocNet/external/selective_search
The above packages are necessary in case you want to try the demo on an image of your choise.
In order to run experiments (e.g. train or test models) on PASCAL VOC datasets you will need to download and prepare the corresponding data. For that purpose you will have to:
- Download the VOC datasets and VOCdevkit:
# VOC2007 DATASET wget http://host.robots.ox.ac.uk/pascal/VOC/voc2007/VOCtrainval_06-Nov-2007.tar # VOC2007 train+val set wget http://host.robots.ox.ac.uk/pascal/VOC/voc2007/VOCtest_06-Nov-2007.tar # VOC2007 test set wget http://host.robots.ox.ac.uk/pascal/VOC/voc2007/VOCdevkit_08-Jun-2007.tar # VOC2007 devkit # VOC2012 DATASET wget http://host.robots.ox.ac.uk/pascal/VOC/voc2012/VOCtrainval_11-May-2012.tar # VOC2012 train+val set wget http://host.robots.ox.ac.uk/pascal/VOC/voc2012/VOCdevkit_18-May-2011.tar # VOC2012 devkit
- Untar the VOC2007 tar files in a directory named
$datasets/VOC2007/VOCdevkit
and the VOC2012 tar files in a directory named$datasets/VOC2012/VOCdevkit
:mkdir $dataset # VOC2007 data: mkdir $datasets/VOC2007 mkdir $datasets/VOC2007/VOCdevkit tar xvf VOCtrainval_06-Nov-2007.tar -C $datasets/VOC2007/VOCdevkit tar xvf VOCtest_06-Nov-2007.tar -C $datasets/VOC2007/VOCdevkit tar xvf VOCdevkit_08-Jun-2007.tar -C $datasets/VOC2007/VOCdevkit # VOC2012 data: mkdir $datasets/VOC2012 mkdir $datasets/VOC2012/VOCdevkit tar xvf VOCtrainval_11-May-2012.tar -C $datasets/VOC2012/VOCdevkit tar xvf VOCdevkit_18-May-2011.tar -C $datasets/VOC2012/VOCdevkit
- They should have the following structure:
# VOC2007 structure: $datasets/VOC2007/VOCdevkit/ # VOC2007 development kit $datasets/VOC2007/VOCdevkit/VOCcode/ # VOC2007 development kit code $datasets/VOC2007/VOCdevkit/VOC2007/ # VOC2007 images, annotations, etc # VOC2012 structure: $datasets/VOC2012/VOCdevkit/ # VOC2012 development kit $datasets/VOC2012/VOCdevkit/VOCcode/ # VOC2012 development kit code $datasets/VOC2012/VOCdevkit/VOC2012/ # VOC2012 images, annotations, etc
- Create symlink of the
$datasets
directory at$LocNet/datasets
:ln -sf $datasets $LocNet/datasets
- Download the pre-computed Edge Box and Selective Search proposals of PASCAL images (click the links) and place them on the following locations:
$LocNet/data/edge_boxes_data # Edge Box proposals $LocNet/data/selective_search_data # Selective Search proposals
To test the object detection pipeline with the (pre-)trained models on VOC2007 test do:
# 1) open matlab from $LocNet directory
cd $LocNet
matlab
# 2) run in the matlab command line enviroment
>> script_test_object_detection_pipeline_PASCAL('VGG16_Reduced_MRCNN','VGG16_LocNet_InOut','bbox_proposals','edge_boxes','gpu_id',1);
The above command will use the Reduced MR-CNN recognition model and the LocNet In-Out localization model located in the following directories:
$LocNet/models-exps/VGG16_Reduced_MRCNN # Reduced MR-CNN recognition model
$LocNet/models-exps/VGG16_LocNet_InOut # LocNet In-Out localization model
In general, you can specify the recognition and localization models that will be tested by giving the directory names of the corresponding models as the 1st and 2nd input arguments in the script_test_object_detection_pipeline_PASCAL
script. For more instructions on how to test the (pre-)trained models see the script:
$LocNet/code/script_test_object_detection_pipeline_commands.m
To train your own LocNet localization models on PASCAL VOC dataset:
- Downlaod the archive file of VGG16-Net model pre-trained on ImageNet classification task and then unzip and untar it in the following location:
$LocNet/data/vgg16_pretrained
. - Then
# 1) open matlab from $LocNet directory cd $LocNet matlab # 2) run in the matlab command line enviroment >> script_train_LocNet_PASCAL('VGG16_LocNet_InOut_model','gpu_id',1,'loc_type','inout');
The above command will train a LocNet In-Out localization model in the union of VOC2007 train+val and VOC2012 train+val datasets using both Selecive Search and Edge Box proposals. The trained model will be saved in the destination directory:
$LocNet/models-exps/VGG16_LocNet_InOut_model
of which the directory name is the string that is given as 1st input argument in the script script_train_LocNet_PASCAL
. For more instructions on how to train your own LocNet localization models see the script:
$LocNet/code/script_train_localization_models_commands.m
Note: for training each of the LocNet models you will require a GPU with at least 6 Gbytes of memory.