This is a python implementation of Selective Search [1][2]. It is forked from belltailjp/selective_search.py to backport it from Python 3 to Python 2.7.
The Selective Search is used as a preprocess of object detection/recognition pipeline.
It finds regions likely to contain any objects from an input image regardless of its scale and location,
that allows detectors to concentrate only for such 'prospective' regions.
Therefore you can configure more computationally efficient detector,
or use more rich feature representation and classification method [3]
compared to the conventional exhaustive search scheme.
For more details about the method, please refer the original paper.
This implementation is based on the journal edition of the original paper, and giving similar parameter variations.
- CMake (>= 3.3.2)
- GCC (>= 4.8.2)
- Python 2.7
- For required packages, see
requirements.txt
- For required packages, see
- Boost (>= 1.58.0) built with python support
- If you get errors building the C++ for selective_search on Mac OS X you should install boost and boost-python via brew, compile them from source, and have them generate universal binaries:
brew install --universal --build-from-source -vd boost
brew install --universal --build-from-source -vd boost-python
- If you get errors building the C++ for selective_search on Mac OS X you should install boost and boost-python via brew, compile them from source, and have them generate universal binaries:
- Boost.NumPy
- If you got an error building on Linux, see belltailjp/Boost.NumPy); if you got an error building on Mac OS X, you probably need to generate a universal binary, see [https://github.com/BradNeuberg/Boost.NumPy])(https://github.com/BradNeuberg/Boost.NumPy) for a Mac OS X-specific fork of Boost.Numpy based on cmake.
This implementation contains some C++ code which wraps the Efficient Graph-Based Image Segmentation [4] tool used for generating an initial value. It works as a python module, so build it first.
% git clone https://github.com/belltailjp/selective_search_py.git
% cd selective_search_py
% wget http://cs.brown.edu/~pff/segment/segment.zip; unzip segment.zip; rm segment.zip
% cmake .
% make
Then you will see a shared object segment.so
in the directory. If you are on Mac OS X you will see segment.dylib
-- you must manually move this over to be segment.so
to work correctly.
Keep it on the same directory of main Python script, or referrable location described in LD_LIBRARY_PATH
on Linux or DYLD_FALLBACK_LIBRARY_PATH
on Mac OS X.
showcandidate demo allows you to interactively see the result of selective search.
% ./demo_showcandidates.py --image image.jpg
You can choose any combination of parameters on the left side of the screen. Then click the "Run" button and wait for a while. You will see the generated regions on the right side.
By changing the slider on the bottom, you can increase/decrease number of region candidates. The more slider goes to left, the more confident regions are shown like this:
showhierarchy demo visualizes colored region images for each step in iteration.
% ./demo_showhierarchy.py image.jpg --k 500 --feature color texture --color rgb
If you want to see labels composited with the input image, give a particular alpha-value.
% ./demo_showhierarchy.py image.jpg --k 500 --feature color texture --color rgb --alpha 0.6
Algorithm of the method is described in Journal edition of the original paper in detail ([1]). For diversification strategy, this implementation supports to vary the following parameter as the original paper proposed.
- Color space
- RGB, Lab, rgI, HSV, normalized RGB and Hue
- C of Color invariance [5] is currently not supported.
- Similarity measure
- Texture, Color, Fill and Size
- Initial segmentation parameter k
- As the initial (fine-grained) segmentation, this implementation uses [4]. k is one of the parameters of the method.
You can give any combinations for each strategy.
If you just want to use this implementation as a black box, only the selective_search
module is necessary to import.
from selective_search import *
img = skimage.io.imread('image.png')
regions = selective_search(img)
for v, (i0, j0, i1, j1) in regions:
...
Then you can get a list regions sorted by score in ascending order. Regions with larger score (latter elements of the list) are considered as 'non-prospective' regions, so they can be filtered out as you need.
To change parameters, just give a list of values for each diversification strategy. Note that they must be given as a list.
selective_search
returns a single list of generated regions which contains every combination of selective search result.
This result is also sorted.
regions = selective_search(img, \
color_spaces = ['rgb', 'hsv'],\ #color space. should be lower case.
ks = [50, 150, 300],\ #k.
feature_masks = [(0, 0, 1, 1)]) #indicates whether S/C/T/F similarity is used, respectively.
This implementation contains automated unit tests using Nose.
To execute the full tests, type:
% nosetests
This implementation is publicly available under the MIT license. See LICENSE.txt for more details.
However regarding the selective search method itself, authors of the original paper have not mention anything so far. Please ask the original authors if you have any concens.
[1] J. R. R. Uijlings et al., Selective Search for Object Recognition, IJCV, 2013
[2] Koen van de Sande et al., Segmentation As Selective Search for Object Recognition, ICCV, 2011
[3] R. Girshick et al., Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation, CVPR, 2014
[4] P. Felzenszwalb et al., Efficient Graph-Based Image Segmentation, IJCV, 2004
[5] J. M. Geusebroek et al., Color invariance, TPAMI, 2001